Re-examining the Suitability of the Raised Floor for Data Center Applications
|
|
|
- Julia Ball
- 10 years ago
- Views:
Transcription
1 Re-examining the Suitability of the Raised Floor for Data Center Applications By Neil Rasmussen White Paper #19 Revision 1
2 Executive Summary The circumstances that gave rise to the development and use of the raised floor in the data center environment are examined. Many of the reasons for the raised floor no longer exist, and the problems associated with raised floors suggest that their widespread use is no longer justified or desirable for many applications. 2
3 Introduction The raised floor is a ubiquitous feature of data centers. In fact, one common definition of a data center is that it is a computing space with a raised floor. The basic science and engineering of the raised floor was fully developed in the 1960s and was described in detail in the 1983 Federal Information Processing Standard 94. The basic design of raised floors has remained relatively unchanged for 40 years. In the telecommunications business the raised floor has never been common. The convergence of telecommunications and IT systems has raised the question of how to decide which approach to use. Recently, more and more IT data centers are being constructed without the use of the raised floor. A review of the history of the raised floor offers insight into this trend. Elements of the Raised Floor The raised floor was developed and implemented as a system intended to provide the following functions: A cold air distribution system for cooling IT equipment Tracks, conduits, or supports for data cabling Conduits for power cabling A copper ground grid for grounding of equipment A location to run chilled water or other utility piping To understand the evolution of the raised floor, it is important to examine each of these functions and what original requirement caused the raised floor to be the appropriate solution. In addition, it is useful to see how the original requirement has changed over time. In the following sections, the original and current requirements relating to the above functions are contrasted. 3
4 A cold air distribution system for cooling IT equipment Original requirement Cool air required near equipment air inlets. Cooling requirement does not change significantly during the lifetime of the data center. Multiple A/C units can feed the system in order to provide fault tolerance. Current requirement Cool air required near equipment air inlets. New IT equipment can operate at 25kW or more per rack requiring 4x the airflow that typical raised floors are designed for; unobstructed underfloor height of 1m is required to attain high density. Cooling requirement changes many times during the lifetime of the data center as equipment moves in and out. Multiple A/C units can feed the system in order to provide fault tolerance. Newer high density, high efficiency air conditioners are closely coupled to the IT cabinets and do not use the raised floor for cool air supply. The raised floor concept meets the original requirement and, due to the ability to change and move vented tiles, it also is well suited for the current requirement. However, the heights of traditional raised floors are inadequate, and many new air conditioner designs simply don t use the floor for the cool air supply. Tracks, co nduits, or supports for data cabling Original requirement Bulky multi-conductor copper data cables connect between cabinets. Cables must be as short as possible to ensure correct operation. Cables need to be hidden from view. Cables are not changed during data center life. Current requirement Thin fiber and copper network cables. Cables are changed many times during data center life. Easy access to cables required. The raised floor was the only practical way to meet the original requirement, but is no longer necessary and is poorly suited to the current requirement due to the difficulty of data cable access. For this reason, most data centers today that use raised floor have some or all data cabling overhead. Conduits for power cabling Original requirement IT equipment is hard-wired with dedicated circuits. Circuit does not change over lifetime of data center. Current requirement IT equipment plugs into standard receptacles. Much higher number of distinct IT equipment per sq. ft. IT Equipment subject to change every 2 years. Branch circuits subject to change multiple times over lifetime of data center. 4
5 The raised floor meets the original and current requirements but has a disadvantage related to access for wiring changes. Overhead power distribution has historically not had any advantages. Power distribution remains a key reason why raised floor is deployed. A copper ground grid for grounding of equipment Original requirement Direct coupled data signal integrity requires interconnected equipment be hard-grounded together to less than 0.1V differential to preserve correct operation and prevent damage. Current requirement Copper network cabling is galvanically transformer isolated and not subject to interference from ground shift up to 1000V and routinely interconnects distances over 50m (160ft). Fiber network cabling is totally immune to ground shift. The need for a copper ground grid has been virtually eliminated. The ground connection between racks and to the branch circuit panels meets the current requirement. A location to run chilled water or other utility piping Original requirement Some IT equipment requires direct cold water piping for cooling. Current requirement IT equipment does not require direct cold water piping for cooling. However, newer row-oriented and rackoriented air conditioning systems do require water or refrigerant piping. This piping must either be under the floor, or, in some designs may be overhead. The raised floor was the only practical way to meet the requirement to deliver water piping to IT equipment. The new requirement for refrigerant or water piping does require either underfloor or overhead installation. If underfloor piping is selected, then a raised floor is required but it only needs to have a height of 16 inches (0.4m) or less. New joint-less overhead piping technology is an alternative that eliminates the need for a raised floor. The Problems of Using a Raised Floor The examination above indicates that the raised floor was a very effective and practical way to meet the original requirements of early data centers. It is also apparent that many of the original requirements that dictate the use of the raised floor no longer exist. In fact, data center requirements have evolved and changed significantly. It is important to review problems that are actually created by the raised floor. Earthquake The raised floor greatly increases the difficulty of assuring or determining a seismic rating for a data center. Supporting equipment above the floor on a grid greatly compromises the ability to anchor equipment. 5
6 Because each installation is different, it is almost impossible to test or validate the seismic rating of an installation. This is a very serious problem in cases where a seismic withstand capability is specified. In and around Kobe Japan, during the great earthquake of 1995, data centers experienced an extraordinary range of earthquake damage. Many data centers which should have been operational within hours or days were down for more than a month when a large number of supposedly earthquake-rated raised floor systems buckled, sending IT equipment crashing through the floor. Damaged equipment needed to be fished out and repaired or replaced in complex and time consuming operations. During the World Trade Center collapse of 2001, nearby data centers which should have survived the tragedy were seriously damaged and experienced extended down time when impacts to the buildings caused raised floor systems to buckle and collapse. A down time of 5 weeks as was typical near Kobe, corresponds to 50,000 minutes as compared with the 5 minutes per year of downtime required to achieve 5-nines reliability. This is 10,000 times worse than the 5- nines design value. If earthquake downtime is considered 10% of the availability budget, then the data centers near Kobe could not achieve 5-nines reliability unless an earthquake of that magnitude were to occur only once every 100,000 years, which would not be a realistic assumption. In areas which are subject to any earthquake activity it is unreasonable to expect 5-nines availability using a raised floor. Even an attempt to do so would be effectively unverifiable. This is one of the reasons why telephone central office facilities do not use raised floors. This is the single most compelling reason why raised floors are no longer desirable for high availability data centers. Access The fact that equipment turnover in a modern data center is around two years gives rise to the situation where data and power cabling is subject to nearly continuous change. The difficulty of access to this cabling when it is under the raised floor results in delay and costs associated with changing needs. Floor loading Typical equipment racks can reach 2000 lbs (907 kg) in weight capacity, and may need to be rolled to be relocated. In addition, the equipment used to move and locate equipment needs data center access. Special reinforcement may be required in a raised floor environment, and in some cases this capability is restricted to certain aisles. Ensuring floor loading requirements are not exceeded requires significant cost and planning. The full load capability of a raised floor is only realized when all of the tiles are in place. The buckling (lateral) strength of the floor depends on the presence of the tiles. However, tiles and even rows of tiles are seen routinely pulled in a data center when frequently required cabling changes or maintenance is performed. This situation can give rise to unexpected catastrophic collapse of the raised floor. 6
7 Headroom In some potential data center locations the loss of headroom resulting from a raised floor is not acceptable. This can limit the options for locating a data center. In Japan, it is common for the floor of the next overhead level of the building to be cut out to create the headroom required. Conduit When cabling is run under a raised floor it becomes subject to special fire regulations. The raised floor is considered under the codes to be an air plenum. Due to the moving and distributed air the fire codes consider a fire in an air plenum to be a special risk. Therefore, cabling under the raised floor is required to be enclosed in fire-rated conduit, which may be metal or a special fire rated polymer. The result is considerable cost and complexity to install this conduit, and a particularly difficult problem when conduit changes are required in an operating data center. Security The raised floor is a space where people or devices may be concealed. In the case of data centers which are partitioned with cages, such as co-location facilities, the raised floor represents a potential way to enter and access caged areas. This is a reason why many co-location facilities do not use raised floor systems. Power distribution The number of branch circuits per square foot in the modern data center is much greater than it was at the time when the raised floor architecture was developed. During the mainframe era, a single hardwired high amperage branch circuit could service a cabinet using 6 floor tiles or 24 square feet (2.2 square meters). Today, this same area could contain two racks, each of which could require 12 kw of 120V circuits with an A and B feed, for a total of 12 branch circuits. The density of the resulting conduits associated with this dramatic increase in branch circuits represent a serious obstacle to underfloor air flow. This can require a raised floor height of 4 feet (1.2 meters) to ensure the needed airflow. Increasing the height of the raised floor compromises the structural integrity and compounds cost, floor loading, and earthquake issues. Cleaning The raised floor is an area which is not convenient to clean. Dust, grit, and various items normally accumulate under the raised floor and are typically abandoned there since the difficulty and accident risk associated with cleaning this area are considered to be serious obstacles. The act of removing a floor tile may cause dramatic shifts in the air motion under the floor, which can and has caused grit or even objects to be blown out into equipment or the eyes of personnel. Safety A tile left open poses a severe and unexpected risk to operators and visitors moving in the data center. In data centers with 4 foot (1.2 meters) or higher raised floors the risk of death resulting from a fall into an open tile location increases greatly. Equipment is frequently moved over the course of the lifetime of today s data centers; this creates the risk that the floor loading will be exceeded leading to a floor collapse. 7
8 Cost The raised floor represents a significant cost. The typical cost of raised floor including engineering, material cost, fabrication, installation, and inspection is on the order of $20 per square foot ($215 per square meter). Furthermore, the maximum space that might be ultimately utilized by the data center is normally built-out with a raised floor whether or not the current, near term, or even the actual ultimate requirement requires the use of this space. The cost of $20 per square foot ($215 per square meter) does not include the extra costs associated with power and data cabling. This is a considerable cost, which should only be incurred if it is actually required. The Barriers to Eliminating the Raised Floor Although more and more installations have eliminated the raised floor, and the benefits of eliminating the raised floor are substantial, some data centers are still designed using a raised floor. Interviews with users of raised floors conducted by APC have identified the following barriers. Perception The raised floor is an icon symbolizing the high availability enterprise data center. For many companies the presentation of their data center is an important part of facility tours for key customers. Data centers that do not have a raised floor are perceived as incomplete or deficient or less than the highest quality. The result is that raised floor is installed as part of creating an image. In some cases, raised floors have been installed that are not used for cooling, wiring and, in fact, have no function at all other than to create the desired image. This issue is by far the largest barrier to the elimination of the raised floor. Cooling design Data center designers and operators value the flexibility that is provided by raised floor cooling designs. The raised floor provides some opportunity to move air vent tiles around in order to achieve a desired temperature profile. This is more difficult in a ducted system using overhead air pipes. In addition, there is much more experience in the design of raised floor air distribution design and designers are therefore better able to predict system performance. However, newer high density, high efficiency air conditioning systems are closely coupled to the IT racks and no longer use the raised floor as a cool air supply, which eliminates this barrier for new designs. Power distribution Due to the shift from a smaller number of larger IT devices to a relatively much larger number of smaller IT devices, the number of branch circuits per square foot in the modern data center is much greater than it was at the time when the raised floor architecture was developed. A location for running these branch circuits is needed. If a raised floor is not used, these circuits must be provided using overhead distribution. Installing and maintaining ceiling mounted branch circuits is perceived to be more difficult than managing underfloor 8
9 branch circuits. However, in most typical data centers, access above the racks is much easier than access under the raised floor. Designing Without a Raised Floor The costs and problems associated with a raised floor can be eliminated only if a practical and available alternative exists. Fortunately, a number of design options allow some flexibility. A complete discussion of these alternatives is beyond the scope of this white paper. In general, the methods for cooling without a raised floor fall into the three categories summarized in the following table. Table 1 Cooling methods without a raised floor Class of Data Center Small Data Centers / Data Rooms (<1000 sq ft) (<93 sq m) Medium Data Centers ( sq ft) ( sq m) Method of Cooling Rack, wall, or ceiling mounted air conditioning units operating without ductwork. Traditional design: Floor standing CRAC units with ductless distribution and with dropped ceiling return plenum. Newer high density, high efficiency design: Row-oriented or rack-oriented cooling systems with overhead joint-less water or refrigerant piping Large Data Centers (>5000 sq ft) (>465 sq m) Traditional design: Floor standing CRAC or roof mount units with ducted overhead cold air distribution combined with open ceiling or dropped ceiling air return. Newer high density, high efficiency design: Row-oriented or rack-oriented cooling systems with overhead jointless water or refrigerant piping These methods have all been used, but the equipment and the design guidelines for these types of installations are less mature when compared with that of raised floor designs. Therefore many systems of traditional design that operate without a raised floor are unique. Engineering firms and suppliers must add to the product base and the knowledge base before these types of installations can be implemented with the predictability offered by the raised floor approach. In general, the traditional methods of implementing cooling in environments without a raised floor exhibit performance that is difficult to predict. The pressures to solve the problems of high density, high efficiency, and predictability have given rise to new air conditioning technologies based on row-oriented and rack-oriented methods. These new air conditioning systems are tightly coupled to the IT loads and integrate into rack cabinets or may be located overhead. Because of the ability of these systems to simultaneously achieve high density, high efficiency, and predictability, their adoption is increasing. One important consequence of these new cooling technologies is that they do not require a raised floor. These technologies are described in more detail in APC White Paper #130, The Advantages of Row and Rack-Oriented Cooling Architectures for Data Centers. 9
10 Recent innovations in overhead rack power distribution have provided a lower cost option when compared with under-floor power distribution. When combined with the emerging overhead cooling methods, it becomes practical to consider eliminating the raised floor, particularly in small and medium sized data centers. Conclusion Many of the reasons that led to the application of the raised floor no longer exist. The absence of a compelling requirement combined with the problems associated with raised floors suggest that their widespread use is no longer economically or technically justified in most applications, particularly in small and medium sized data centers. Recently introduced solutions have solved the technical barriers to the elimination of the raised floor. In fact, the highest density and highest efficiency data center power and cooling technologies specifically do not require a raised floor. Nevertheless, data centers are likely to use raised floors for some time due to large base of experience with raised floor design, traditions of locating piping and wiring under the floor, and the intangible issues of perception and image. About the Author: Neil Rasmussen is the Chief Technical Officer of APC-MGE. He establishes the technology direction for the world s largest R&D budget devoted to power, cooling, and rack infrastructure for critical networks. Neil is currently leading the effort at APC to develop high-efficiency, modular, scalable data center infrastructure solutions and is the principal architect of the APC InfraStruXure system. Prior to founding APC in 1981, Neil received his Bachelors and Masters degrees from MIT in electrical engineering where he did his thesis on the analysis of a 200MW power supply for a tokamak fusion reactor. From 1979 to 1981, he worked at MIT Lincoln Laboratories on flywheel energy storage systems and solar electric power systems. 10
Raised Floors vs Hard Floors for Data Center Applications
Raised Floors vs Hard Floors for Data Center Applications White Paper 19 Revision 3 by Neil Rasmussen Executive summary Raised floors were once a standard feature of data centers, but over time a steadily
A Scalable, Reconfigurable, and Efficient Data Center Power Distribution Architecture
A Scalable, Reconfigurable, and Efficient Data Center Power Distribution Architecture By Neil Rasmussen White Paper #129 Executive Summary Significant improvements in efficiency, power density, power monitoring,
Efficiency and Other Benefits of 208 Volt Over 120 Volt Input for IT Equipment
Efficiency and Other Benefits of 208 Volt Over 120 Volt Input for IT Equipment By Neil Rasmussen White Paper #27 Revision 2 Executive Summary Decisions made regarding the distribution of 208V or 120V power
How To Calculate Total Cooling Requirements For Data Center
Calculating Total Cooling Requirements for Data Centers White Paper 25 Revision 3 by Neil Rasmussen > Executive summary This document describes how to estimate heat output from information technology (IT)
Strategies for Deploying Blade Servers in Existing Data Centers
Strategies for Deploying Blade Servers in Existing Data Centers By Neil Rasmussen White Paper #125 Revision 1 Executive Summary When blade servers are densely packed, they can exceed the power and cooling
A Scalable, Reconfigurable, and Efficient Data Center Power Distribution Architecture
A Scalable, Reconfigurable, and Efficient Data Center Power Distribution Architecture White Paper 129 Revision 1 by Neil Rasmussen > Executive summary Significant improvements in efficiency, power density,
Calculating Total Cooling Requirements for Data Centers
Calculating Total Cooling Requirements for Data Centers By Neil Rasmussen White Paper #25 Revision 2 Executive Summary This document describes how to estimate heat output from Information Technology equipment
Avoiding Costs from Oversizing Data Center and Network Room Infrastructure
Avoiding Costs from Oversizing Data Center and Network Room Infrastructure White Paper 37 Revision 6 by Neil Rasmussen > Executive summary The physical and power infrastructure of data centers and network
Increasing Data Center Efficiency by Using Improved High Density Power Distribution
Increasing Data Center Efficiency by Using Improved High Density Power Distribution By Neil Rasmussen White Paper #128 Executive Summary A new approach to power distribution for high density server installations
Improving Rack Cooling Performance Using Airflow Management Blanking Panels
Improving Rack Cooling Performance Using Airflow Management Blanking Panels By Neil Rasmussen White Paper #44 Revision 3 Executive Summary Unused vertical space in open frame racks and rack enclosures
Avoiding Costs from Oversizing Data Center and Network Room Infrastructure
Avoiding Costs from Oversizing Data Center and Network Room Infrastructure White Paper 37 Revision 7 by Neil Rasmussen > Executive summary The physical infrastructure of data centers and network rooms
Electrical Efficiency Modeling for Data Centers
Electrical Efficiency Modeling for Data Centers By Neil Rasmussen White Paper #113 Revision 1 Executive Summary Conventional models for estimating electrical efficiency of data centers are grossly inaccurate
Avoidable Mistakes that Compromise Cooling Performance in Data Centers and Network Rooms
Avoidable Mistakes that Compromise Cooling Performance in Data Centers and Network Rooms By Neil Rasmussen White Paper #49 Executive Summary Avoidable mistakes that are routinely made when installing cooling
Improving Rack Cooling Performance Using Airflow Management Blanking Panels
Improving Rack Cooling Performance Using Airflow Management Blanking Panels White Paper 44 Revision 4 by Neil Rasmussen > Executive summary Unused vertical space in open frame racks and rack enclosures
Electrical Efficiency Modeling of Data Centers
Electrical Efficiency Modeling of Data Centers By Neil Rasmussen White Paper 113 Executive Summary The conventional models used to estimate electrical efficiency of data centers are grossly inaccurate
APC APPLICATION NOTE #112
#112 Best Practices for Deploying the InfraStruXure InRow SC By David Roden Abstract The InfraStruXure InRow SC (ACSC100 and ACSC101) is a self-contained air conditioner for server rooms and wiring closets.
Implementing Energy Efficient Data Centers
Implementing Energy Efficient Data Centers By Neil Rasmussen White Paper #114 Executive Summary Electricity usage costs have become an increasing fraction of the total cost of ownership (TCO) for data
Power and Cooling Capacity Management for Data Centers
Power and Cooling Capacity for Data Centers By Neil Rasmussen White Paper #150 Executive Summary High density IT equipment stresses the power density capability of modern data centers. Installation and
AC vs DC Power Distribution for Data Centers
AC vs DC Power Distribution for Data Centers By Neil Rasmussen White Paper #63 Revision 4 Executive Summary Various types of DC power distribution are examined as alternatives to AC distribution for data
Determining Total Cost of Ownership for Data Center and Network Room Infrastructure
Determining Total Cost of Ownership for Data Center and Network Room Infrastructure White Paper #6 Revision 3 Executive Summary An improved method for measuring Total Cost of Ownership of data center and
The Advantages of Row and Rack- Oriented Cooling Architectures for Data Centers
The Advantages of Row and Rack- Oriented Cooling Architectures for Data Centers By Kevin Dunlap Neil Rasmussen White Paper #130 Executive Summary Room cooling is an ineffective approach for next-generation
Benefits of. Air Flow Management. Data Center
Benefits of Passive Air Flow Management in the Data Center Learning Objectives At the end of this program, participants will be able to: Readily identify if opportunities i where networking equipment
Power and Cooling Guidelines for Deploying IT in Colocation Data Centers
Power and Cooling Guidelines for Deploying IT in Colocation Data Centers White Paper 173 Revision 0 by Paul Lin and Victor Avelar Executive summary Some prospective colocation data center tenants view
Overload Protection in a Dual-Corded Data Center Environment
Overload Protection in a Dual-Corded Data Center Environment White Paper 206 Revision 0 by Neil Rasmussen Executive summary In a dual-corded environment, the loss of power on one path will cause the load
Use of the Signal Reference Grid in Data Centers
Use of the Signal Reference Grid in Data Centers By Neil Rasmussen White Paper #87 Executive Summary Signal reference grids are automatically specified and installed in data centers despite the fact that
Selecting an Industry-Standard Metric for Data Center Efficiency
Selecting an Industry-Standard Metric for Data Center Efficiency By Neil Rasmussen White Paper #157 Executive Summary The development of standards for the measurement and specification of data center efficiency
Guidelines for Specification of Data Center Power Density
Guidelines for Specification of Data Center Power Density By Neil Rasmussen White Paper #120 Executive Summary Conventional methods for specifying data center density are ambiguous and misleading. Describing
IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT
DATA CENTER RESOURCES WHITE PAPER IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT BY: STEVE HAMBRUCH EXECUTIVE SUMMARY Data centers have experienced explosive growth in the last decade.
Cooling Strategies for IT Wiring Closets and Small Rooms
Cooling Strategies for IT Wiring Closets and Small Rooms White Paper 68 Revision 1 by Neil Rasmussen and Brian Standley > Executive summary Cooling for IT wiring closets is rarely planned and typically
Cooling Strategies for IT Wiring Closets and Small Rooms
Cooling Strategies for IT Wiring Closets and Small Rooms By Neil Rasmussen Brian Standley White Paper #68 Executive Summary Cooling for IT wiring closets is rarely planned and typically only implemented
Creating Order from Chaos in Data Centers and Server Rooms
Creating from in Data Centers and Server Rooms By Dennis Bouley White Paper #119 Executive Summary Data center professionals can rid themselves of messy racks, sub-standard under floor air distribution,
Raised Floor Data Centers Prepared for the Future
Raised Floor Data Centers Prepared for the Future Raised Floors Protect My Investment by Providing a Foundation for Current and Future Technologies. Creating the Future Proof Data Center Environment The
Cooling Audit for Identifying Potential Cooling Problems in Data Centers
Cooling Audit for Identifying Potential Cooling Problems in Data Centers By Kevin Dunlap White Paper #40 Revision 2 Executive Summary The compaction of information technology equipment and simultaneous
Guidelines for Specification of Data Center Power Density
Guidelines for Specification of Data Center Power Density White Paper 120 Revision 1 by Neil Rasmussen > Executive summary Conventional methods for specifying data center density are ambiguous and misleading.
Power and Cooling for Ultra-High Density Racks and Blade Servers
Power and Cooling for Ultra-High Density Racks and Blade Servers White Paper #46 Introduction The Problem Average rack in a typical data center is under 2 kw Dense deployment of blade servers (10-20 kw
Choosing Close-Coupled IT Cooling Solutions
W H I T E P A P E R Choosing Close-Coupled IT Cooling Solutions Smart Strategies for Small to Mid-Size Data Centers Executive Summary As high-density IT equipment becomes the new normal, the amount of
APC APPLICATION NOTE #92
#92 Best Practices for Designing Data Centers with the InfraStruXure InRow RC By John Niemann Abstract The InfraStruXure InRow RC is designed to provide cooling at the row and rack level of a data center
The Different Types of UPS Systems
The Different Types of UPS Systems By Neil Rasmussen White Paper #1 Revision 5 Executive Summary There is much confusion in the marketplace about the different types of UPS systems and their characteristics.
High-Efficiency AC Power Distribution for Data Centers
High-Efficiency AC Power Distribution for Data Centers White Paper 128 Revision 2 by Neil Rasmussen > Executive summary The use of 240 volt power distribution for data centers saves floor space, simplifies
Calculating Space and Power Density Requirements for Data Centers
Calculating Space and Power Density Requirements for Data Centers White Paper 155 Revision 0 by Neil Rasmussen Executive summary The historic method of specifying data center power density using a single
Creating Order from Chaos in Data Centers and Server Rooms
Creating from in Data Centers and Server Rooms White Paper 119 Revision 1 by Dennis Bouley > Executive summary Data center professionals can rid themselves of messy racks, sub-standard under floor air
Alan Matzka, P.E., Senior Mechanical Engineer Bradford Consulting Engineers
Information Technology and Data Centers HVAC Showcase November 26, 2013 Alan Matzka, P.E., Senior Mechanical Engineer Bradford Consulting Engineers Welcome. Today s webinar is being recorded and will be
AisleLok Modular Containment vs. Legacy Containment: A Comparative CFD Study of IT Inlet Temperatures and Fan Energy Savings
WH I TE PAPE R AisleLok Modular Containment vs. : A Comparative CFD Study of IT Inlet Temperatures and Fan Energy Savings By Bruce Long, Upsite Technologies, Inc. Lars Strong, P.E., Upsite Technologies,
Rittal Liquid Cooling Series
Rittal Liquid Cooling Series by Herb Villa White Paper 04 Copyright 2006 All rights reserved. Rittal GmbH & Co. KG Auf dem Stützelberg D-35745 Herborn Phone +49(0)2772 / 505-0 Fax +49(0)2772/505-2319 www.rittal.de
Managing Cooling Capacity & Redundancy In Data Centers Today
Managing Cooling Capacity & Redundancy In Data Centers Today About AdaptivCOOL 15+ Years Thermal & Airflow Expertise Global Presence U.S., India, Japan, China Standards & Compliances: ISO 9001:2008 RoHS
Data Center Airflow Management Retrofit
FEDERAL ENERGY MANAGEMENT PROGRAM Data Center Airflow Management Retrofit Technology Case Study Bulletin: September 2010 Figure 1 Figure 2 Figure 1: Data center CFD model of return airflow short circuit
Data Center Projects: Establishing a Floor Plan
Data Center Projects: Establishing a Floor Plan White Paper 144 Revision 2 by Neil Rasmussen and Wendy Torell > Executive summary A floor plan strongly affects the power density capability and electrical
The Different Types of UPS Systems
White Paper 1 Revision 7 by Neil Rasmussen > Executive summary There is much confusion in the marketplace about the different types of UPS systems and their characteristics. Each of these UPS types is
Greening Commercial Data Centres
Greening Commercial Data Centres Fresh air cooling giving a PUE of 1.2 in a colocation environment Greater efficiency and greater resilience Adjustable Overhead Supply allows variation in rack cooling
Data Center Rack Systems Key to Business-Critical Continuity. A White Paper from the Experts in Business-Critical Continuity TM
Data Center Rack Systems Key to Business-Critical Continuity A White Paper from the Experts in Business-Critical Continuity TM Executive Summary At one time, data center rack enclosures and related equipment
- White Paper - Data Centre Cooling. Best Practice
- White Paper - Data Centre Cooling Best Practice Release 2, April 2008 Contents INTRODUCTION... 3 1. AIR FLOW LEAKAGE... 3 2. PERFORATED TILES: NUMBER AND OPENING FACTOR... 4 3. PERFORATED TILES: WITH
The Different Types of UPS Systems
Systems White Paper 1 Revision 6 by Neil Rasmussen Contents Click on a section to jump to it > Executive summary There is much confusion in the marketplace about the different types of UPS systems and
Best Practices for Wire-free Environmental Monitoring in the Data Center
White Paper 11800 Ridge Parkway Broomfiled, CO 80021 1-800-638-2638 http://www.42u.com [email protected] Best Practices for Wire-free Environmental Monitoring in the Data Center Introduction Monitoring for
Optimizing Network Performance through PASSIVE AIR FLOW MANAGEMENT IN THE DATA CENTER
Optimizing Network Performance through PASSIVE AIR FLOW MANAGEMENT IN THE DATA CENTER Lylette Macdonald, RCDD Legrand Ortronics BICSI Baltimore 2011 Agenda: Discuss passive thermal management at the Rack
Green Data Centre Design
Green Data Centre Design A Holistic Approach Stantec Consulting Ltd. Aleks Milojkovic, P.Eng., RCDD, LEED AP Tommy Chiu, EIT, RCDD, LEED AP STANDARDS ENERGY EQUIPMENT MATERIALS EXAMPLES CONCLUSION STANDARDS
BRUNS-PAK Presents MARK S. EVANKO, Principal
BRUNS-PAK Presents MARK S. EVANKO, Principal Data Centers of the Future and the Impact of High Density Computing on Facility Infrastructures - Trends, Air-Flow, Green/LEED, Cost, and Schedule Considerations
How To Run A Data Center Efficiently
A White Paper from the Experts in Business-Critical Continuity TM Data Center Cooling Assessments What They Can Do for You Executive Summary Managing data centers and IT facilities is becoming increasingly
An Improved Architecture for High-Efficiency, High-Density Data Centers
An Improved Architecture for High-Efficiency, High-Density Data Centers By Neil Rasmussen White Paper #126 Executive Summary Data center power and cooling infrastructure worldwide wastes more than 60,000,000
in other campus buildings and at remote campus locations. This outage would include Internet access at the main campus.
DATA CENTER The current Campus Data Center is located within historic Founders Hall which is in a less-than ideal location for a variety of reasons, but is typical of an existing campus IT infrastructure
The Different Types of Air Conditioning Equipment for IT Environments
The Different Types of Air Conditioning Equipment for IT Environments By Tony Evans White Paper #59 Executive Summary Cooling equipment for an IT environment can be implemented in 10 basic configurations.
Element D Services Heating, Ventilating, and Air Conditioning
PART 1 - GENERAL 1.01 OVERVIEW A. This section supplements Design Guideline Element D3041 on air handling distribution with specific criteria for projects involving design of a Data Center spaces B. Refer
Analysis of the UNH Data Center Using CFD Modeling
Applied Math Modeling White Paper Analysis of the UNH Data Center Using CFD Modeling By Jamie Bemis, Dana Etherington, and Mike Osienski, Department of Mechanical Engineering, University of New Hampshire,
Data Centers and Mission Critical Facilities Operations Procedures
Planning & Facilities Data Centers and Mission Critical Facilities Operations Procedures Attachment A (Referenced in UW Information Technology Data Centers and Mission Critical Facilities Operations Policy)
Our data centres have been awarded with ISO 27001:2005 standard for security management and ISO 9001:2008 standard for business quality management.
AIMS is Malaysia and South East Asia s leading carrier neutral data centre operator and managed services provider. We provide international class data storage and ancillary services, augmented by an unrivaled
Specialty Environment Design Mission Critical Facilities
Brian M. Medina PE Associate Brett M. Griffin PE, LEED AP Vice President Environmental Systems Design, Inc. Mission Critical Facilities Specialty Environment Design Mission Critical Facilities March 25,
How to Install Modular Wiring, Cabling and Air Distribution Devices beneath a Access Floor without a Fire Suppression System
How to Install Modular Wiring, Cabling and Air Distribution Devices beneath a Access Floor without a Fire Suppression System July 28, 2003 We are pleased to provide the following joint statement by CII,
Calculating Total Power Requirements for Data Centers
Calculating Total Power Requirements for Data Centers By Richard Sawyer White Paper #3 Executive Summary Part of data center planning and design is to align the power and cooling requirements of the IT
Unified Physical Infrastructure (UPI) Strategies for Thermal Management
Unified Physical Infrastructure (UPI) Strategies for Thermal Management The Importance of Air Sealing Grommets to Improving Smart www.panduit.com WP-04 August 2008 Introduction One of the core issues affecting
OCTOBER 2010. Layer Zero, the Infrastructure Layer, and High-Performance Data Centers
Layer Zero, the Infrastructure Layer, and The demands on data centers and networks today are growing dramatically, with no end in sight. Virtualization is increasing the load on servers. Connection speeds
OPERATOR - FACILITY ADDRESS - DATE
. Facility Criteria (RFP).0 Location Characteristics Criteria Requirements 1 Ownership and Management: Please identify the owner(s) and managers of the property and indicate if there is on-site management.
Our data centres have been awarded with ISO 27001:2005 standard for security management and ISO 9001:2008 standard for business quality management.
AIMS is Malaysia and South East Asia s leading carrier neutral data centre operator and managed services provider. We provide international class data storage and ancillary services, augmented by an unrivaled
Verizon SMARTS Data Center Design Phase 1 Conceptual Study Report Ms. Leah Zabarenko Verizon Business 2606A Carsins Run Road Aberdeen, MD 21001
Verizon SMARTS Data Center Design Phase 1 Conceptual Study Report Ms. Leah Zabarenko Verizon Business 2606A Carsins Run Road Aberdeen, MD 21001 Presented by: Liberty Engineering, LLP 1609 Connecticut Avenue
Data Center Components Overview
Data Center Components Overview Power Power Outside Transformer Takes grid power and transforms it from 113KV to 480V Utility (grid) power Supply of high voltage power to the Data Center Electrical Room
Comparing Data Center Power Distribution Architectures
Comparing Data Center Power Distribution Architectures White Paper 129 Revision 3 by Neil Rasmussen and Wendy Torell Executive summary Significant improvements in efficiency, power density, power monitoring,
Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings
WHITE PAPER Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings By Lars Strong, P.E., Upsite Technologies, Inc. Kenneth G. Brill, Upsite Technologies, Inc. 505.798.0200
Cooling Small Server Rooms Can Be. - Jim Magallanes Computer Room Uptime: www.cruptime.com Uptime Racks: www.uptimeracks.com
Cooling Small Server Rooms Can Be Inexpensive, Efficient and Easy - Jim Magallanes Computer Room Uptime: www.cruptime.com Uptime Racks: www.uptimeracks.com Server Rooms Description & Heat Problem Trends
Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED.
Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED. Presentation Goals & Outline Power Density Where we have been- where we are now - where we are going Limitations of Air Cooling
AT&T Internet Data Center Site Specification Chicago Area (Lisle, IL)
AT&T Internet Data Center Site Specification Chicago Area (Lisle, IL) Site Biography Market Area: Chicago, IL Location: Lisle, IL 60532 Address: 4513 Western Avenue CLLI,POP, NPA/NXX Code: LSLEILAA, EQLAILAA
High Density Data Centers Fraught with Peril. Richard A. Greco, Principal EYP Mission Critical Facilities, Inc.
High Density Data Centers Fraught with Peril Richard A. Greco, Principal EYP Mission Critical Facilities, Inc. Microprocessors Trends Reprinted with the permission of The Uptime Institute from a white
In Row Cooling Options for High Density IT Applications
In Row Cooling Options for High Density IT Applications By Ramzi Y. Namek, PE, LEED AP, BD+C; Director of Engineering 1 Users wishing to deploy high density IT racks have several In Row Cooling solutions
Cooling a hot issue? MINKELS COLD CORRIDOR TM SOLUTION
Cooling a hot issue? MINKELS COLD CORRIDOR TM SOLUTION The heat load in data centers is increasing rapidly, the adaptation of a cooling strategy is often necessary. Many data centers and server rooms use
Doc Ref: 2015.06.01.002. 602.783.1112 baselayer.com 2015 BASELAYER TECHNOLOGY, LLC
TM 602.783.1112 baselayer.com 2015 BASELAYER TECHNOLOGY, LLC TM D400E Data Module (60Hz) The EDGE D400E is designed to meet high power density requirements for both the HPC and Web Scale environments while
The Optical Fiber Ribbon Solution for the 10G to 40/100G Migration
The Optical Fiber Ribbon Solution for the 10G to 40/100G Migration Bill Charuk, SR Product Manager, Data Center Solutions Introduction With the advent of data centers in communication infrastructure, various
The New Data Center Cooling Paradigm The Tiered Approach
Product Footprint - Heat Density Trends The New Data Center Cooling Paradigm The Tiered Approach Lennart Ståhl Amdahl, Cisco, Compaq, Cray, Dell, EMC, HP, IBM, Intel, Lucent, Motorola, Nokia, Nortel, Sun,
Data Centers. Mapping Cisco Nexus, Catalyst, and MDS Logical Architectures into PANDUIT Physical Layer Infrastructure Solutions
Data Centers Mapping Cisco Nexus, Catalyst, and MDS Logical Architectures into PANDUIT Physical Layer Infrastructure Solutions 1 Introduction The growing speed and footprint of data centers is challenging
Data center upgrade proposal. (phase one)
Data center upgrade proposal (phase one) Executive Summary Great Lakes began a recent dialogue with a customer regarding current operations and the potential for performance improvement within the The
Containment Solutions
Containment Solutions Improve cooling efficiency Separate hot and cold air streams Configured to meet density requirements Suitable for commercial and SCEC endorsed cabinets Available for retrofit to SRA
Energy-efficient & scalable data center infrastructure design
Seminar am 30. September 2010 in Wetzlar Nachhaltigkeit durch UPI (Unified Physical Infrastructure) Vortrag von Stefan Fammler: Energieeffiziente Strategien im Rechenzentrum Energy-efficient & scalable
NeuStar Ultra Services Physical Security Overview
NeuStar Ultra Services Physical Security Overview Mission-critical network operations must have the highest standards of quality, security and reliability. NeuStar Ultra Services currently locates its
Education Evolution: Scalable Server Rooms George Lantouris Client Relationship Manager (Education) May 2009
Education Evolution: Scalable Server Rooms George Lantouris Client Relationship Manager (Education) May 2009 Agenda Overview - Network Critical Physical Infrastructure Cooling issues in the Server Room
Enclosure and Airflow Management Solution
Enclosure and Airflow Management Solution CFD Analysis Report Des Plaines, Illinois December 22, 2013 Matt Koukl- DECP Mission Critical Systems Affiliated Engineers, Inc. Contents Executive Summary...
AT&T Internet Data Center Site Specification Washington-DC Area (Ashburn, VA)
AT&T Internet Data Center Site Specification Washington-DC Area (Ashburn, VA) Site Biography Market Area: Washington, DC Area Location: Ashburn, VA 20147-6011 Address: All deliveries should be sent to
LEVEL 3 DATA CENTER ASSESSMENT
Nationwide Services Corporate Headquarters 410 Forest Street Marlborough, MA 01752 USA Tel: 800-342-5332 Fax: 508-303-0579 www.eecnet.com LEVEL 3 DATA CENTER ASSESSMENT Submitted by: Electronic Environments
Our data centres have been awarded with ISO 27001:2005 standard for security management and ISO 9001:2008 standard for business quality management.
AIMS is Malaysia and South East Asia s leading carrier neutral data centre operator and managed services provider. We provide international class data storage and ancillary services, augmented by an unrivaled
Portable Energy Efficient Data Centre Solution
Portable Energy Efficient Data Centre Solution Workspace Technology Limited Technology House, PO BOX 11578, Sutton Coldfield West Midlands B72 1ZB Telephone: 0121 354 4894 Facsimile: 0121 354 6447 www.workspace-technology.com
Reducing Room-Level Bypass Airflow Creates Opportunities to Improve Cooling Capacity and Operating Costs
WHITE PAPER Reducing Room-Level Bypass Airflow Creates Opportunities to Improve Cooling Capacity and Operating Costs By Lars Strong, P.E., Upsite Technologies, Inc. 505.798.000 upsite.com Reducing Room-Level
