Direct Contact Liquid Cooling (DCLC ) PRODUCT GUIDE. Industry leading data center solutions for HPC, Cloud and Enterprise markets.



Similar documents
The New Data Center Cooling Paradigm The Tiered Approach

Rittal Liquid Cooling Series

Data Center Equipment Power Trends

Bytes and BTUs: Holistic Approaches to Data Center Energy Efficiency. Steve Hammond NREL

GREEN FIELD DATA CENTER DESIGN WATER COOLING FOR MAXIMUM EFFICIENCY. Shlomo Novotny, Vice President and Chief Technology Officer, Vette Corp.

Enabling an agile Data Centre in a (Fr)agile market

Data Center Rack Level Cooling Utilizing Water-Cooled, Passive Rear Door Heat Exchangers (RDHx) as a Cost Effective Alternative to CRAH Air Cooling

HPC TCO: Cooling and Computer Room Efficiency

Data Center Facility Basics

Liquid Cooling Solutions for DATA CENTERS - R.M.IYENGAR BLUESTAR LIMITED.

Energy Efficient Data Centre at Imperial College. M. Okan Kibaroglu IT Production Services Manager Imperial College London.

THE WORLD S MOST EFFICIENT DATA CENTER

Oil Submersion Cooling for Todayʼs Data Centers. An analysis of the technology and its business implications

Airflow Simulation Solves Data Centre Cooling Problem

Dealing with Thermal Issues in Data Center Universal Aisle Containment

Data Center 2020: Delivering high density in the Data Center; efficiently and reliably

High-density modular data center making cloud easier. Frank Zheng Huawei

Blade Server & Data Room Cooling Specialists

How To Design A Data Center

Managing Data Centre Heat Issues

Elements of Energy Efficiency in Data Centre Cooling Architecture

Virginia Tech. Background. Case Summary. Results

7 Best Practices for Increasing Efficiency, Availability and Capacity. XXXX XXXXXXXX Liebert North America

AIRAH Presentation April 30 th 2014

Data Sheet FUJITSU Server PRIMERGY CX400 M1 Multi-Node Server Enclosure

Energy Efficiency Opportunities in Federal High Performance Computing Data Centers

Specialty Environment Design Mission Critical Facilities

PRIMERGY server-based High Performance Computing solutions

Optimization of energy consumption in HPC centers: Energy Efficiency Project of Castile and Leon Supercomputing Center - FCSCL

How To Power And Cool A Data Center

Data Sheet Fujitsu PRIMERGY CX1000 S1 with 38 cloud server nodes

Data Center Technology: Physical Infrastructure

DataCenter 2020: first results for energy-optimization at existing data centers

Energy Recovery Systems for the Efficient Cooling of Data Centers using Absorption Chillers and Renewable Energy Resources

Capgemini UK Infrastructure Outsourcing

ICT and the Green Data Centre

Green Data Center and Virtualization Reducing Power by Up to 50% Pages 10

Datasheet Fujitsu PRIMERGY CX1000 S1 with 38 CX120 S1 The Cool-Central Architecture

European Code of Conduct for Data Centre. Presentation of the Awards 2014

SmartDesign Intelligent, Integrated Infrastructure For The Data Center. Solutions For Business-Critical Continuity

Raised Floor Data Centers Prepared for the Future

DataCenter Data Center Management and Efficiency at Its Best. OpenFlow/SDN in Data Centers for Energy Conservation.

BRUNS-PAK Presents MARK S. EVANKO, Principal

Rack Solutions Integration Solution Service

IT White Paper MANAGING EXTREME HEAT: COOLING STRATEGIES FOR HIGH-DENSITY SYSTEMS

University of St Andrews. Energy Efficient Data Centre Cooling

- White Paper - Data Centre Cooling. Best Practice

In Row Cooling Options for High Density IT Applications

Challenges In Intelligent Management Of Power And Cooling Towards Sustainable Data Centre

European Business Reliance Centres when Commodity serves Green IT

How to Build a Data Centre Cooling Budget. Ian Cathcart

CURBING THE COST OF DATA CENTER COOLING. Charles B. Kensky, PE, LEED AP BD+C, CEA Executive Vice President Bala Consulting Engineers

Environmental Data Center Management and Monitoring

Building a data center. C R Srinivasan Tata Communications

Typical Air Cooled Data Centre Compared to Iceotope s Liquid Cooled Solution

Overview of Green Energy Strategies and Techniques for Modern Data Centers

Intel Cluster Ready Appro Xtreme-X Computers with Mellanox QDR Infiniband

opdc Cut your data center OPEX upto 70%

Huawei Data Center Facility Integration Solution

Excool Indirect Adiabatic and Evaporative Data Centre Cooling. World s Leading Indirect Adiabatic and Evaporative Data Centre Cooling

How High Temperature Data Centers & Intel Technologies save Energy, Money, Water and Greenhouse Gas Emissions

Free Cooling in Data Centers. John Speck, RCDD, DCDC JFC Solutions

Cooling a hot issue? MINKELS COLD CORRIDOR TM SOLUTION

2006 APC corporation. Cooling Solutions and Selling Strategies for Wiring Closets and Small IT Rooms

IMPROVING DATA CENTER EFFICIENCY AND CAPACITY WITH AISLE CONTAINMENT

Data Sheet FUJITSU Server PRIMERGY CX420 S1 Out-of-the-box Dual Node Cluster Server

Cloud Computing Data Centers

Trends in Data Center Design ASHRAE Leads the Way to Large Energy Savings

Leveraging Renewable Energy in Data Centers: Present and Future

An Introduction to Cold Aisle Containment Systems in the Data Centre

Energy Efficiency in New and Existing Data Centers-Where the Opportunities May Lie

Datacenter Design Issues

Paul Oliver Sales Director. Paul Oliver, Airedale Tom Absalom, JCA

Cooling Capacity Factor (CCF) Reveals Stranded Capacity and Data Center Cost Savings

NetApp Data Center Design

Data Center Components Overview

Transmission System Total Cost of Ownership

GREEN DATA CENTER DESIGN

How High Temperature Data Centers and Intel Technologies Decrease Operating Costs

Reducing Data Center Energy Consumption

Data Center Consolidation Trends & Solutions. Bob Miller Vice President, Global Solutions Sales Emerson Network Power / Liebert

The Mission Critical Data Center Understand Complexity Improve Performance

How High Temperature Data Centers & Intel Technologies save Energy, Money, Water and Greenhouse Gas Emissions

Insourcing vs. Outsourcing. Bart Leung Senior Associate, Ehvert Mission Critical

Data Centre Stockholm II, Sweden Flexible, advanced and efficient by design.

(Examples of various colocation sites)

Datasheet Fujitsu PRIMERGY CX1000 S1 with 38 cloud server nodes The Cool-Central Architecture

Implementing Energy Efficient Data Centers

Reference Guide: SmartDesign Approach: Common Data Center Challenges:

Transcription:

Direct Contact Liquid Cooling (DCLC ) PRODUCT GUIDE Industry leading data center solutions for HPC, Cloud and Enterprise markets.

Leadership, Collaboration & Expertise CooIIT Systems Innovates and Delivers CoolIT Systems is a world leader in direct contact liquid cooling for the Data Center and Desktop markets. CoolIT has been dedicated to the invention and design of safe and reliable liquid cooling technology since 2001. As an experienced innovator with 44 patents and more than 1.8 million liquid cooling units deployed in desktop computers, servers and data centres around the world, CoolIT s Rack DCLC technology is the top choice for many OEMs and system integrators. Rack DCLC is a three module approach (Server, Manifold and Heat Exchanger) that allows for a tremendous amount of product flexibility when integrating liquid cooling. With a reputation for working intimately with customers, CoolIT Systems delivers industry leading liquid cooling solutions for a variety of applications. With options for data centers with and without facility water hook up, any server in any rack can be liquid cooled with CoolIT s hardware, and benefit from immediate and measurable CAPEX and OPEX benefits. Proven Technology Selling 30-40K DCLC units/month >1.85M units sold worldwide 99.998% leak free and improving 45 issued patents, 19 more in process Awards Profit500 2014 #34 in Canada 2014 (1780% 5-year growth rate) Geoff Lyon Nominated for E&Y Entrepreneur of the Year Deloitte 2014 Technology Fast 50 Canada - ranked #23 Deloitte 2014 Technology Fast 500 North America - ranked #257 We strive to deliver the highest level of product and service so that when customers choose CoolIT, they know they are getting the best. Geoff Lyon, CEO/CTO CoolIT Technology Partners OEM Partners 2009-2014, CoolIT Systems. All Rights Reserved

Focused on Your Cooling Challenges Liquid Cooling for the Modern Data Center Power savings, increased rack density and high performance are the three key advantages of using liquid cooling. Performance Efficiency Density Rack DCLC facilitates peak performance for higher powered or overclocked processors. Rack DCLC provides a significant reduction in total data center energy consumed. Rack DCLC allows for densities of 45kW or greater per rack. Immediate CAPEX & OPEX Benefits Modern air conditioning techniques are very expensive and inefficient at cooling computers. Data centers often employ chiller-based air systems that consume 50% of all data center power. W3-W5 warm liquid cooling eliminates the need for chilled water supply and delivers a rapid ROI (within 1-6 months). Trust is a huge factor when putting liquids in close proximity to critical electronics. CoolIT s technology outfitted with Stäubli s non-spill connectors works flawlessly. Skeptics become converts. Rusty Robertson, Aerospace/Defense Application Specialist Staubli Corporation CAPEX CAPEX (annualized) (annualized) OPEX OPEX (annualized) (annualized) Rack Rack Power Power Total IT Total Power IT Power 1.5 MW 1.5 MW 1.5 MW 1.5 MW Total Data Total Center Data Center Power Power 2.7 MW 2.7 MW *Including *Including Cooling IT Cooling UPS & IT AUX UPS & AUX 1.8 MW 1.8 MW # of Racks # of Racks Required Required 100 100 55 55 IT Density IT Density 15kW/rack 15kW/rack 27.7kW/rack 27.7kW/rack * Using * Koomey s Using Koomey s data center data costing center costing methods methods WWW.COOLITSYSTEMS.COM T: 403.235.4895 Toll Free: 1.866.621.2665

A Modular Approach to Liquid Cooling Cooling Solutions for All IT Demands Rack DCLC Modules are designed for a flexible fit to various compute environments. While Server Modules and Manifold Modules are installed with each system and are local to the rack, the appropriate heat rejection method may vary. Rack DCLC offers a variety of heat exchanging modules depending on load requirements and availability of facility water. Server Module Patented cold plate technology supports multiple combinations of CPUs/GPUs and other types of interposers, heat pipes, or heat spreaders. Manifold Module Reliable stainless steel manifolds employing dry break, quick connect technology in horizontal, multiple horizontal, or vertical chassis, blades or any other server array. Heat Exchange Module Your choice of heat exchange module. Liquid-to-Air or Liquid-to-Liquid. UNLESS OTHERWISE SPECIFIED: DIMENSIONS ARE IN MM(INCH) TOLERANCES: DATE: (MM/DD/YY) ANGULAR: ± 0.5 ± 0.2(0.01) X.X (X.XX) = ± 0.05(0.001) X.XX (X.XXX) PROPRIETARY AND CONFIDENTIAL THIS DRAWING IS THE PROPERTY OF COOLIT SYSTEMS AND IS SUBJECT TO RETURN UPON REQUEST. THE DRAWING Requires Facility Water CHx and AHx solutions support all sizes of racks. Various sizing and designs are available depending on customer needs and environmental factors. Modules Server Module Manifold Module CHx650 Module CHx40 Module AHx35 Module AHx20 Module 2009-2014, CoolIT Systems. All Rights Reserved No Facility Water SCALE: NTS COOLIT SYSTEMS INC. 3920-29th Street NE Tel: (403)235-4895 Calgary, AB T1Y 6B6 Fax: (403)770-8306 AHx Module, L109cm x W58cm x H13cm

Liquid-to-liquid heat exchange that manages a network of IT cabinets. The CHx650 is designed to manage the distribution of clean, treated coolant to and from a network of IT cabinets. This CHx accepts W3-W5 warm facility water and can manage 650kW of processor load per network. Key Features Features Supports 40kW+ cooling capacity per rack W3-W5 warm water cooling Redundant pumps Full control system Benefits Key Benefits Dramatically increase CPU/GPU density Eliminates the need for chillers and CRACs Monitor system health remotely ROI in 1 year or less The CHx650 Module can be customized to fit various data center environments. Standard equipment groups offer N+1 redundancy and when used in tandem can provide tier 4 resiliency. CoolIT s Rack DCLC solutions help us achieve what some would consider the impossible. Patrick Scateni, VP Sales & Marketing CIARA CHx650 Cooling Capacity Maximum Cooling Load (kw) vs. Facility Liquid Input Temperature ( C) Capacity (kw) 1000 750 500 250 Cooling Capacity (@ 30 C facility water) Power Consumption (Max) Racks Per Solution ROI (months) 650kW 4.2kW 5-20 1-6 0 10 20 30 40 50 Facility Water Integration Yes Facility Input Temperature ( C) WWW.COOLITSYSTEMS.COM T: 403.235.4895 Toll Free: 1.866.621.2665

Liquid-to-liquid heat exchange that manages a single IT cabinet. The CHx40 is designed to manage the distribution of clean, treated coolant within a single IT cabinet. This 2U CHx accepts W3-W5 warm facility water and can manage 40kW of processor load per rack. Features Key Features Supports 40kW+ cooling capacity per rack W3-W5 warm water cooling Redundant pumps 2U form factor Full control system Key Benefits CHx40 Cooling Capacity Dramatically increase CPU/GPU density Eliminates the need for chillers and CRACs Monitor system health remotely ROI in 1 year or less Cooling Capacity (kw) 60 50 40 30 20 10 0 Facility water temperature 20 C 30 C 40 C 10 20 30 40 Cooling Capacity (@ 30 C facility water) Power Consumption (Max) Racks Per Solution ROI (months) Facility Water Integration 40kW 95 watts 1 1-6 Yes Facility Flow Rate (L/min) CASE STUDY: Cooling a Supercomputer Intel Selects CoolIT s Rack DCLC AHx35 to Liquid Cool Top500 Cherry Creek Cluster CoolIT Systems teamed up with Intel and Supermicro to create an innovative HPC cluster that ranked #400 on the Top500 list of supercomputers and an impressive #41 on the Green 500 list. The cluster uses CoolIT s Rack DCLC AHx solution and Supermicro s ultra-dense FatTwin server platform, featuring 9936 cores with a peak performance of 131.2 teraflops. The system consumed a total of 74.25kW of power even though each Supermicro FatTwin node featured dual Intel Xeon processors and three Intel Xeon Phi coprocessors that were generating over 1,000 watts of heat per node. Cherry Creek Cluster running live on the SC13 floor. 2009-2014, CoolIT Systems. All Rights Reserved

Liquid-to-air heat exchange for faster, more powerful clusters in less space. AHx Module - various designs available (hot aisle, cold aisle, chimney, or other) AHx35 Cooling Capacity *Cherry Creek AHx Cooling Capacity (kw) 60 50 40 30 20 10 0 10 20 30 40 50 Facility Air Inlet Temperature ( C) The AHx35 is a unique top-of-rack solution that can manage 35kW of processor load without the need of facility water. The AHx dissipates heat to the surrounding environment via a liquid-to-air heat exchanger. Key Features Features Supports 35kW+ cooling capacity per rack No facility water required Redundant pumps Full control system Benefits Key Benefits Dramatically increase CPU/GPU density Stand alone rack-based liquid cooling Monitor system health remotely ROI in 1 year or less Cooling Capacity (@ 25 C air) Power Consumption (Max) Racks Per Solution ROI (months) Facility Water Integration 35kW 795 watts 1 1-12 No Liquid-to-air heat exchange. Key Features Supports 20kW+ cooling capacity per rack No facility water required Redundant pumps Full control system Key Benefits Dramatically increase CPU/GPU density Stand alone rack-based liquid cooling Monitor system health remotely ROI in 1 year or less The AHx20 packs an incredible amount of heat load management into a small form factor. This top-of-rack solution provides superior processor performance and very high densities. The AHx20 is perfect for compute environments that are short on space, big on compute, and don t have access to facility water. Cooling Capacity (@ 25 C air) Power Consumption (Max) Racks Per Solution ROI (months) Facility Water Integration 20kW 300 watts 1 1-12 No WWW.COOLITSYSTEMS.COM T: 403.235.4895 Toll Free: 1.866.621.2665

@coolitsystems sales@coolitsystems.com www.coolitsystems.com