Vette Girls Corvette Association Online Magazine February/March 2015
Vette LiquiCool TM Solution Arlene Allen University of California Santa Barbara Director,...
-
date post
19-Dec-2015 -
Category
Documents
-
view
213 -
download
0
Transcript of Vette LiquiCool TM Solution Arlene Allen University of California Santa Barbara Director,...
Vette LiquiCoolTM Solution
Arlene AllenUniversity of California Santa BarbaraDirector, Information Systems & Computing
Rob PerryExecutive Manager
1
Data Center Trends - Staggering Energy Consumption and Cost of Energy
3Source: EPA 2007 Report to Congress
Energy unit price has increased an average of 4% YOY in the USA and 11% YOY GloballyData Center energy consumption is growing by 12% annually
4
Source: Belady, C., “In the Data Center, Power and Cooling Costs More than IT Equipment it Supports”, Electronics Cooling Magazine (Feb 2007)
Data Center facility costs are growing 20% vs. IT spend of 6% Operating costs over lifetime of a server ~ 4X original purchase cost Cooling infrastructure can consume up to 55% of Data Center energy
Data Center Trends – Operating Expense Exceeds Capital Expense in less than 1
year
UCSB – “The Problem”
UCSB’s existing Data Center is being renovated for research computing and is forcing the corporate/miscellaneous IT equipment into a new space.This new space is not designed to be a Data Center. The footprint is small, the power is limited by existing building wiring and using traditional air-cooling topology is not feasible.The new space limitations requires the load density to increase from a typical density of 6kW or less to a higher density of 10-16kW per rack
6
LiquiCool - “The Solution”
7
Move the corporate/miscellaneous IT racks into the new space
Tap into the existing building chilled water system
Install Vette LiquiCool Rear Door Heat Exchangers on every rack
Install Vette LiquiCool Coolant Distribution Unit for a secondary loop
Install rack-mount UPS in every rack
LiquiCool - “The Solution”
LiquiCool™ – A complete cooling solution for the consolidation and scale-out of compute infrastructure in today’s sustainable Data Centers
8
Reduces white space requirements by more than 55%
Cuts cooling energy consumption by 50% or more when compared to traditional air-cooled Data Centers
Allows 8X the amount of compute power in a typical IT enclosure
Lowers carbon footprint by more 50% or more vs. air-cooling
Bottom Line: Payback in less than 1 year when compared to traditional computer room air-conditioning
Front of Enclosure
Rear of Enclosure
LiquiCool - How does it work?
9
Based on IBM IP & Technology Licenses (>30 years of water cooling experience)Rear Door Heat Exchanger (RDHx) replaces existing rear door of IT enclosureRDHx has chilled water Supply & Return quick connections at bottom OR topRaised floor becomes optionalChilled water circulates through tube+fin coil from Supply connectionEquipment exhaust air passes through coil and is cooled before re-entering the roomFin + tube
Heat exchanger
Cold Supply Water
Heated Water
LiquiCool System
10
Passive RDHx provides 100% sensible coolingNo condensation, no need for reheat or humidification
CDU creates a fully isolated, temperature controlled Secondary LoopChilled water source - city water, building chilled water, packaged chiller…
1010
Temperature:7oC / 45o FWater pressure:100-200 psi
Temperature:10-17oC 50-63oF
Water pressure:30-70 psi
RDHx - External View
11Top Feed Connections Bottom Feed Connections
PassiveNo electrical connectionsNo moving parts
No FansNo powerNo noise
Attaches to rearNo need to rearrange racksDoes not consume valuable floor space, adds 4-6” to rear
Close-coupledNeutralizes at the source
RDHx - Internal View
12Tube & Fin coil
Air-bleed valves
Bottom Feed HoseConnections and drain valve
Protective barrier
Thermal Image - Before & After100% Heat Neutralization
13
RDHx Cooling in Action
Rear Door Heat Exchanger Door openedServer Leaving Temp 102ºF (38.9ºC)
RDHx reduces Leaving Temperature by 28ºF (15.4ºC)! 14
Rear Door Heat Exchanger Door closedServer Leaving Temp: 74ºF (23.5ºC)
Temperature readings taken in the rear of a fully populated Enclosure
15
Industry Standard Enclosure
Remove existing rack rear door & hinges
RDHx is Compatible with most major IT Enclosures
Mount Transition Frame (if needed)
• Max. Cooling Capacity: 33kW • Coolant: Chilled Water (above dew
point)• Dimensions: 76.6“ H x 4.6“ D x 23.6“ W
(1945mm x 117mm x 600mm)
• Weight – empty: 63lbs (29kg)• Liquid Volume: 1.5 Gallons (5.7 Liters)• Liquid Flow Rate: 6-10 GPM (23-38 L/min)• Head Loss: 7 psi (48 kPa) at 10 GPM (38
L/min)• System Input Power: None required• Noise: None• Couplings: Drip-free stainless steel
quick- connects• Connection Location: Bottom or Top Feed 16
RDHx General Specifications
• Water to water heat exchanger with pumps, controls and chilled water valve• Creates an isolated secondary cooling loop
– 100% sensible cooling, no condensation– Small water volume (tens of gallons)– Easier to control water quality
• Redundant, fault-tolerant design • 120kW or 150kW capacity
– Supports 6-12 RDHx– Optional internal manifold for quick expansion
• SNMP & ModBus communications
17
Coolant Distribution Unit (CDU)
• Power Consumption: 2.6 kW• Pump Capacity: 63 GPM at 30psi (240 L/min at 207
kPa)• Primary Head Loss: 10.2 psi at 63 GPM (70 kPa at
240 L/min)• Minimum Approach Temperature (100% load):
• 120kW unit - 12°F (6.7 °C)• 150kW unit - 8°F (4.4 °C)• 63 GPM (240 L/min) on primary and secondary
CDU Simplified
18
Floor-mount CDU Internal - Front
Controller
Redundant variable speed pumps
Redundant valves
Reservoir tank
Inverter drive
Brazed plate heat exchanger
Casters and Drain 19
Floor-mount CDU Internal - Rear
Primary side water filter
Optional Secondary Loop Distribution Manifold
Primary supply and return connections
Optional Secondary Loop Flex Tails
20
21
Hose Kits & External Manifolds
Connects to flex tails on CDU secondary sideISO B or Sweated ConnectionsStandard & custom configurations
Each Vette Hose Kit consists of a flexible Supply hose and a Return hoseFactory assembled and tested to IBM specifications and standardsQuick-connect drip-free couplings on one end OR both endsStraight hoses for raised floor environments, right angle hoses for non-raised floor environmentsStandard lengths from 3ft. to 50ft.
Water Treatment
22
CORROSION
MICROBIO
FOU
LINGSC
ALE
Potential Effects of Non-Treatment• Loss of heat transfer• Reduced system efficiency• Reduced equipment life• Equipment failures or leaks• De-ionized water without inhibitors
is corrosive!
Treatment of Cooling Water
Scenario I – Out of Space
23
56% Recovery of
White Space!
Add RDHx – Double your load per rack
Eliminate CRAC units
Scenario II – Out of Power/Capacity
24
Add RDHx
Remove (2) CRAC units
Reduces cooling energy
consumption to free up
capacity for growth
Scenario III – High Density
CRAC units can typically provide efficient environmental control for rack densities
of up to 5kw per rack
25
Adding RDHx allows 8X the
ComputePower!
Reference Sites
26
National Center for HPC, Taiwan
Warwick University, Coventry, UK
Georgia Tech Super Computing Facility - 12 racks at ~24kW each
Front view Rear view
Reference Sites
27
Silicon Valley Leadership Group Case Study - Modular Cooling Systems
SVLG “Chill Off” Results
Vette’s LiquiCool™ solution led the field in cooling capacity and in cooling efficiency!
Vette
LiquiCool - Conclusive Savings for Energy, Space & Cost
Largest % of Data Center OPEX growth is power & cooling relatedCost of energy for cooling is a large (and growing) cost component
Data Center consolidation, virtualization and advanced hardware technology are driving higher power densities per rack and associated white space constraints
Traditional air-cooling is less likely feasible
Purchasing decisions can no longer be made solely on CAPEX TCO must not only be considered, but is core
Value Summary:Reduces white space requirements by more than 55%Cuts cooling energy consumption by 50% or more when compared to traditional air-cooled Data CentersAllows 8X the amount of compute power in a typical IT enclosureLowers carbon footprint by more 50% or more vs. air-coolingBottom Line: Payback in less than 1 year when compared to traditional computer room air-conditioning
30
EndThank You
31