hvac for data centers

59
HVAC for Data Centers Sponsored by: Join the discussion about this Webcast on Twitter at #CSEdatacenterHVAC

Upload: consultingspecifyingengineer

Post on 14-Jan-2015

1.207 views

Category:

Education


6 download

DESCRIPTION

When developing data center energy-use estimations, engineers must account for all sources of energy use in the facility. Most energy consumption is obvious: computers, cooling plant and related equipment, lighting, and other miscellaneous electrical loads. Designing efficient and effective data centers is a top priority for consulting engineers. Cooling is a large portion of data center energy use, second only to the IT load. Although there are several options to help maximize HVAC efficiency and minimize energy consumption, data centers come in many shapes, sizes, and configurations. By developing a deep understanding of their client’s data center HVAC requirements, consulting engineers can help maintain the necessary availability level of mission critical applications while reducing energy consumption.

TRANSCRIPT

Page 1: HVAC for Data Centers

HVAC for Data Centers

Sponsored by:

Join the discussion about this Webcast on Twitter at #CSEdatacenterHVAC

Page 2: HVAC for Data Centers

Today’s Webcast Sponsors:

Page 3: HVAC for Data Centers

Learning Objectives:

1.The audience will learn about codes and guidelines, such as ASHRAE 90.1: Energy Standard for Buildings Except Low-Rise Residential Buildings, and U.S. Green Building Council LEED v4

2.Attendees will learn the relationships between HVAC efficiency and power usage effectiveness (PUE)

3.Viewers will understand the advantages and drawbacks of using an elevated IT equipment inlet temperature

4.Viewers will learn how running IT equipment at partial load affects data center energy efficiency.

Page 4: HVAC for Data Centers

Bill Kosik, PE, CEM, BEMP, LEED AP BD+C,HP Enterprise Business, Technology Services,Chicago, Ill.

Tom R. Squillo, PE, LEED AP, Environmental Systems Design Inc., Chicago, Ill.

Moderator: Jack Smith, Consulting-Specifying Engineer and Pure Power, CFE Media, LLC

Presenters:

Page 5: HVAC for Data Centers

Energy Code Requirements for Data Centers

Tom R. Squillo, PE, LEED AP, Environmental Systems Design

Inc., Chicago, Ill.

Page 6: HVAC for Data Centers

Energy Code Requirements for Data CentersInternational Energy Conservation Code: IECC

• Adopted by eight states and many local jurisdictions

• Written in enforceable code language

ASHRAE Energy Standard for Buildings: ASHRAE 90.1

• Standard instead of code. Now written in enforceable language so it can be adopted locally

• Has more language specific to data centers

California Building Energy Efficiency Standards: Title 24

Page 7: HVAC for Data Centers

Energy Code Requirements for Data CentersWhere Do They Apply?

• Check local jurisdiction for specific requirements or city energy codes

• Many local jurisdictions refer to state codes

• IECC allows compliance with ASHRAE 90.1 instead. This may be an advantage in some instances

• Title 24 compliance required in California

Page 8: HVAC for Data Centers

Current Energy Code Adoption Status (U.S. DOE)

Page 9: HVAC for Data Centers

Projected Energy Code Adoption by end of 2015 (U.S. DOE)

Page 10: HVAC for Data Centers

Energy Code Requirements for Data CentersIECC – 2012:• IECC delineates between simple

and complex systems. Stand alone DX ac units may fall under the simple category. Only very small units under 33,000 Btu/h capacity do not require economizers

• All cooling systems with some form of common piping distribution fall under the complex category

• All complex systems require economizers

Page 11: HVAC for Data Centers

Energy Code Requirements for Data CentersIECC – 2012:C403.4.1 Economizers. Economizers shall comply with Sections C403.4.1.1 through C403.4.1.4• This section requires either an air or water economizer

C403.4.1.1 Design capacity. Water economizer systems shall be capable of cooling supply air by indirect evaporation and providing up to 100% of the expected system cooling load at outdoor air temperatures of 50 F dry bulb/45 F wet bulb and below.• Exception for small systems below 33,000 Btu/h• Unlike ASHRAE 90.1, IECC has no specific exceptions for

data centers that allow lower dry-bulb/wet-bulb temperatures. Dry-coolers are not allowed.

Page 12: HVAC for Data Centers

Energy Code Requirements for Data CentersASHRAE 90.1-2010

• Data centers, considered “process cooling,” were excluded from the requirements of ASHRAE 90.1-2007

• The 2010 version eliminates the “process cooling” exemption, and adds specific language for computer rooms

• To comply with the IECC-2012 code, ASHRAE 90.1-2010 may be used instead.

Page 13: HVAC for Data Centers

Energy Code Requirements for Data Centers

ASHRAE 90.1 – 2010:6.5.1 Economizers. Each Cooling System that has a fan shall include either an air or water economizer meeting the requirements of Sections 6.5.1.1 through 6.5.1.4.

For data centers, economizers are not required for:• Small fan-cooling units less than 135,000 Btu/hr or 65,000 Btu/hr,

depending on climate zone• Extremely hot and humid climate zones• Buildings with no central CHW plant, in which the total computer room

cooling capacity is less than 250 tons• Buildings with a central CHW plant, and the computer room cooling load

is less than 50 tons • Where cooling towers are not allowed• Addition of less than 50-ton computer room capacity to existing building• Various essential facilities (national defense, emergency response, etc.)

Page 14: HVAC for Data Centers

Energy Code Requirements for Data Centers

ASHRAE 90.1 – 2010:6.5.1.2 Water Economizers

6.5.1.2.1 Design capacity. Water economizer systems shall be capable of cooling supply air by indirect evaporation and by providing up to 100% of the expected system cooling load at outdoor air temperatures of 50 F dry bulb/45 F wet bulb and below.• For data centers, the requirement is relaxed slightly to allow

100% economizer cooling at 40 F dry bulb/35 F wet bulb and below

• The code also allows dry-coolers for data centers, but they must provide 100% economizer cooling at 35 F dry bulb

Page 15: HVAC for Data Centers

Energy Code Requirements for Data CentersImportant Changes to ASHRAE 90.1-2013:6.5.1.2 Water Economizers• For data centers, the outdoor temperature limits for 100% water

side economization are not a single condition, but are based on the individual climate zones

6.5.1.6 Economizer Humidification Impact. Systems with hydronic cooling and humidification systems designed to maintain inside humidity at a dew-point temperature greater than 35 F shall use a water economizer if an economizer is required by Section 6.5.1.• This essentially bans air side economizer systems for most

data center systems if using a prescriptive approach.

Page 16: HVAC for Data Centers

Energy Code Requirements for Data CentersImportant Changes to ASHRAE 90.1-2013:6.6 Alternative Compliance Path• For data centers, the HVAC systems can comply by meeting

minimum PUE requirements instead of Section 6.5-Prescriptive Path.

• The minimum PUE values are based on the climate zone and range from 1.30 to 1.61.

• PUE calculation is based on Green Grid Recommendation document dated May, 2011.

• Option 1: Use peak PUE calculation (at 50% and 100% IT load)• Option 2: Use annual PUE, calculated with an approved hourly

energy analysis program (DOE, BLAST, EnergyPlus, etc.)

Page 17: HVAC for Data Centers

Energy Code Requirements for Data CentersTitle 24 – 2013: Highlights Specific to Data Centers• Data centers are exempt from normal

economizer requirements• Air or water economizer required• Air economizer must provide 100%

economization at 55 F dry bulb• Water economizer must provide 100%

economization at 40 F dry bulb/35 F wet bulb

• Economizer exceptions exist for small systems

• Nonadiabatic humidification (steam, infrared) is prohibited

Page 18: HVAC for Data Centers

Energy Code Requirements for Data CentersTitle 24 – 2013: Highlights Specific to Data Centers• Variable speed supply fans required

for DX systems over 5 tons and all CHW systems

• Supply fans shall vary airflow rate as a function of actual load

• Containment required for data centers with a design load exceeding 175 W/sq ft

• Containment exception for expansions and racks below 1 W/sq ft

• Chilled water plants can have no more than 300 tons of air-cooled chillers

Page 19: HVAC for Data Centers

Relationships Between HVAC Efficiency and Power Usage

Effectiveness (PUE)

Bill Kosik, PE, CEM, BEMP, LEED AP BD+C,HP Enterprise Business, Technology

Services,Chicago, Ill.

Page 20: HVAC for Data Centers

• Extreme regional variations in CO2 from electricity generation

• Determine appropriate balance of water and electricity usage

• Climate WILL impact HVAC energy use – select sites carefully

• Use evaporative cooling where appropriate

• Economizer strategy will be driven from climate characteristics

• Design power and cooling modularity to match IT growth

• Plan for power-aware computing equipment• Use aisle containment or direct-cooled cabinets• Design in ability to monitor and optimize PUE in real

time

• Push for highest supply temperatures and lowest moisture levels

• Identify tipping point of server fan energy/inlet temperature

• Minimize data center footprint by using high-density architecture

Dat

a C

ente

rC

limat

eS

yner

gie

sC

on

verg

en

ce

Levels of Optimization

Page 21: HVAC for Data Centers

Air Cooled

Chiller

coupled to chilled

water coil in air

handling unit

(AHU)

Direct

Expansion

packaged in ahu

or separate DX

coil & remote

condenser

Water Cooled

Chiller

coupled with chw coil

in ahu. typical with

open CT & flat plate

HX

Air Cooled

Chiller

coupled to chw

coil in ahu.

typical with

closed CT

Interior AHU

direct outside

air with direct

evaporative

cooling

Exterior AHU

indirect outside

air and indirect

evaporative

cooling

CRAH

Unit

perimeter

air deliver

with chilled

water coil

In-Row Unit

close-coupled in

rack containment

system with

module fans and

chw coil

Rear Door

HX

Individual

rack door chw

HX. Passive

system with

no fan

Overhead

Coil

Module chw

coils. Passive

system with

no fan

Typical Data Center Cooling StrategiesAir Side Economizers Water Side Economizers

System 1 – DEC

Direct Outside Air

Economizer with

Direct Evaporative

Cooling

System 2 – IEC

Recirculating

(Closed) Air System

with Indirect

Evaporative Cooling

System 3 – IDEC

Recirculating (Closed) and

Direct Outside Air System 2

Stage Indirect-Direct

Evaporative Cooling

System 4 – IOA+EC

Indirect Air to Air HX

with Direct

Evaporative Cooling

in Secondary Air

System 5 –

OCT+FP HX

Direct (Open)

Evaporative Cooling

Tower with Flat Plate

System 6 – CCT

w/Spray

Indirect (Closed)

Cooling Tower with

Spray

Mechanical Cooling Options Mechanical Cooling Options

Cooling Configuration OptionsCooling Configuration Options

Page 22: HVAC for Data Centers

0.35 difference in PUE based on climate and cooling system type

PUE Varies with Climate

Page 23: HVAC for Data Centers

Impacts of Climate on Economization Strategy

This analysis shows the percent of total ton-hours that require mechanical cooling. The graphs depict two systems with two water temperatures, 12°C and 18°C:1. Air-cooled chiller with dry-cooled economization2. Air-cooled chiller with evaporative-cooler economization

Page 24: HVAC for Data Centers

The indirect evaporative and indirect air cooling systems have the lowest compressor energy used to cool the facility. The air-cooled DX and air-cooled chiller systems have the highest compressor energy. The air-cooled chiller with economizer is in the middle of the other options.

Impacts of Climate on Economization Strategy

Santiago, CHL

Page 25: HVAC for Data Centers

HVAC System and PUE

Five HVAC options are shown. Each option was analyzed using the same input parameters such as climate attributes, air and water temperatures, etc. Each system performs differently based on the inherent strategies used to cool the data center and provide the proper airflow. For each option, the annual HVAC consumption and annual PUE us shown.

Page 26: HVAC for Data Centers

HVAC System and PUETwo options are shown. The only difference between the two options is the location of the data centers. Everything else, including power, cooling, and ancillary systems are modelled identically. Month by month PUE values are shown as well as monthly HVAC, electrical losses, lighting and other electrical energy. The energy consumption of the data center located in Singapore is markedly higher based on the hot and humid climate.

Page 27: HVAC for Data Centers

Elevated IT Equipment Temperatures

Tom R. Squillo, PE, LEED AP, Environmental Systems Design

Inc., Chicago, Ill.

Page 28: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures Legacy Data Center Design

• Data center supply air set to 50 F to 55 F

• DX systems cycled on/off and fought each other, with little capacity control or communication

• Chilled water temperatures of 40 F to 45 F

• No containment

• Wide variation of temperatures entering IT equipment

Page 29: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures Why Were Low Supply Temperatures Needed?

• Design needed to take into account massive mixing of hot air with supply air

• Temperature of air entering IT equipment at tops of racks still acceptable

• Cold room allowed some ride-through if cooling failed

Why Was This Bad?

• Wastes Energy

– Too much airflow (low delta T)– Inefficient chiller operation– Limited economizer use– Unnecessary dehumidification

• Hot spots

• Inconsistent temperature control

• Inconsistent humidity control

Source: 42u.com

Page 30: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

ASHRAE Thermal Guidelines

• Recommended range for IT inlet conditions

– Temperature: 64 F to 80.6 F

– Humidity: 41.9 F to 59 F dew point or 60% RH

• Extended range for other classes of IT equipment

Source: 42u.com

Page 31: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures Advantages of Elevated Temperatures

• Increased equipment efficiency

– 1.5% to 2.5% increase in chiller efficiency per degree increase in chilled water temperature

– Increasing CHW supply temperature from 50 F to 60 F decreases chiller energy by up to 25%

– Actual increase depends on chiller type and selection

• Decreased unwanted dehumidification at air handling units

– Coil temperature never gets below dew point if CHW temperature is raised

– Eliminates condensate removal issues

• Additional economizer hours

– Actual advantage highly dependent on climate and system type

– Longer equipment life

Page 32: HVAC for Data Centers

Air Side Economizer System: Phoenix, 60 F SA Temperature

• Economization available for 5,038 hr/yr

• Chillers off when outside air temperature is below 60 F (1,910 hr)

Page 33: HVAC for Data Centers

• Economization available for 7,396 hr/yr

• Huge gain in economizer hours due to dry climate

• Chillers off when outside air temperature is below 75 F (4,294 hr)

Air Side Economizer System: Phoenix, 75 F SA Temperature

Page 34: HVAC for Data Centers

Air Side Economizer System: Charlotte, 60 F SA Temperature

• Economization available for 5,300 hr/yr

• More hours of full economization (3,778 hr) than Phoenix

Page 35: HVAC for Data Centers

Air Side Economizer System: Charlotte, 75 F SA Temperature

• Economization available for 5,630 hr/yr

• Due to humid climate, increase in economizer hours is minimal

• Chillers off when outside air temperature is below 75 F (5,300 hr)

Page 36: HVAC for Data Centers

Water Side Economizer System: Phoenix, 60 F SA Temperature

• Economization available for 3,829 hr/yr

• Chillers off when outside air wet bulb temperature is below 33 F (55 hr)

Page 37: HVAC for Data Centers

Water Side Economizer System: Phoenix, 75 F SA Temperature

• Economization available for 8,629 hr/yr

• Huge gain in economizer hours due to dry climate

• Chillers off when outside wet bulb temperature is below 53 F (4,420 hr)

Page 38: HVAC for Data Centers

Water Side Economizer System: Charlotte, 60 F SA Temperature

• Economization available for 4,174 hr/yr

• Chillers off when outside wet bulb temperature is below 33 F (1,009 hr)

Page 39: HVAC for Data Centers

Water Side Economizer System: Charlotte, 75 F SA Temperature

• Economization available for 8,334 hr/yr

• water side economizer has huge increase in economizer hours because less hours are locked out due to OA humidity

• Chillers off when outside wet bulb temperature is below 53 F (4,513 hr)

Page 40: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Bin Data Energy Use Analysis

• Design Criteria– Typical enterprise/co-location data center load

• 10,000 sq ft• 200 W/sq ft• 2 MW of total IT load

– Within ASHRAE recommended conditions• Supply temperature = 75 F• Return temperature = 95 F• Space dew point temperature between 42 F and 59 F• Efficient adiabatic humidification used for analysis

Page 41: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Two Systems and two climates analyzed and compared:

System Options:1. Direct Outside Air Economizer

• Multiple 500 kW capacity rooftop supply/exhaust AHUs• OA Economizer control• Air-cooled chiller system for supplemental cooling

2. Water-cooled Chillers with Cooling Towers• High efficiency variable speed chillers (0.4 kW/ton)• Induced draft cooling towers• Plate and frame heat exchangers in series w/chillers• CRAH units with high efficiency EC fans on raised floor

Climates:3. Phoenix – Hot and dry4. Charlotte – Warm and humid

Page 42: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

60°F Supply Air 65°F Supply Air 70°F Supply Air 75°F Supply Air

-

1,000,000

2,000,000

3,000,000

4,000,000

5,000,000

6,000,000

MPUE1.20

1.30

1.17

1.26

1.15

1.22

1.13

Phoenix Energy Consumption (kWh)Chiller CHW PumpCW Pump AHU FanTower Fan Supply FanExhaust Fan Humidification

1.34

Page 43: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

Direct Outdoor

Air

WC Chillers/CT

60°F Supply Air 65°F Supply Air 70°F Supply Air 75°F Supply Air

-

1,000,000

2,000,000

3,000,000

4,000,000

5,000,000

6,000,000

MPUE1.24

1.19

1.22

1.17

1.20

1.15

1.19

1.13

Charlotte Energy Consumption (kWh)Chiller CHW PumpCW Pump AHU FanTower Fan Supply FanExhaust Fan Humidification

Page 44: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Disadvantages of elevated temperatures

• Working conditions in hot aisle

– Hot aisle temperatures may rise above 100 F in some cases

– OSHA requirements may come into effect

– Think about temporary spot cooling for technology workers

• Temperature ratings of cables and sprinkler heads in hot aisle

– Some cabling rated for 40 C (104 F)

• Reduced ride-through time during cooling failures

– Critical server temperatures can be reached in minutes or seconds in some cases

– Good containment will help reduce hot air recirculation, though may starve servers if system airflow is interrupted

Page 45: HVAC for Data Centers

Elevated IT Equipment Inlet Temperatures

Conclusions:• Increasing IT inlet temperatures can help reduce overall energy

use substantially by:– Increasing chiller efficiency (10 degree rise can increase efficiency

up to 25%)– Reduce humidification requirements– Huge increases in economizer hours

• Be careful of very high temperature conditions in the hot aisles affecting worker comfort and equipment ratings

• Advantages highly dependent on climate and system type– Look at the psych chart for economizer and lock-out hours– air side and water side economizer systems will be affected

differently

Page 46: HVAC for Data Centers

Partial Loads

Bill Kosik, PE, CEM, BEMP, LEED AP BD+C,HP Enterprise Business, Technology

Services,Chicago, Ill.

Page 47: HVAC for Data Centers

PUE Sensitivity to Low IT Loads

How running IT equipment at partial load affects data center energy efficiency.

Page 48: HVAC for Data Centers

• Multiple systems allow for growth without over-provisioning

• Modularity lowers fan energy and increases compressor effectiveness

• Modularity is not the same as spreading the load across multiple pieces of equipment

Efficiency Through Modularity

Page 49: HVAC for Data Centers

Electrical losses will increase as the IT load decreases. This increase must be included in cooling load at different loading points.

Efficiency Through Modularity

Page 50: HVAC for Data Centers

Electrical System Topology and System Efficiency

Page 51: HVAC for Data Centers

As the cooling load is distributed over an increasing number of chillers, the overall power (and energy) grows. To maintain the highest efficiency, the chillers should be run as close as possible to their peak efficiency point.

Page 52: HVAC for Data Centers

Servers Are More Efficient but Use More PowerTrends in Server Turn-Down Ratio

Page 53: HVAC for Data Centers

Server Modularity

45 hot-plug cartridges

Compute, Storage, or Combinationx86 , ARM, or Accelerator

• Single-server = 45 servers per chassis

• Quad-server =180 servers per chassis (future capability) – that is 1800 servers per cabinet or 45 kW

Dual low-latency switches

• Switch Module (45 x 1 GB downlinks)

Compute, Storage, or combination x86 , ARM, or Accelerator

Page 54: HVAC for Data Centers

The PUE values are predicted using data center energy use simulation techniques. Many assumptions are made which affect the predicted energy use and PUE. These ranges are meant to be indicators of the PUE envelope that might be expected based on sub-system efficiency levels, geographic location and methods of operations. Detailed energy use simulation is required to develop more granular and accurate analyses.

Input data for "high" PUE caseSingapore, SGP

Use Water Economizer NOUse Adiabatic Cooling NOLighting (w/SF) 1.50Misc Power (% of IT) 6.0%Electrical System Loss (%) 10.0%Air-cooled evap temp (°F) 65.0Fan Pressure 2.0

Input data for "low" PUE caseHelsinki, FIN

Use Water Economizer YESUse Adiabatic Cooling YESLighting (w/SF) 1.00Misc Power (% of IT) 4.0%Electrical System Loss (%) 8.5%Air-cooled evap temp (°F) 65.0Fan Pressure 2.0

Page 55: HVAC for Data Centers

Codes and Standards References from Today’s Webcast

• HVAC: ASHRAE 90.1: Energy Standard for Buildings Except Low-Rise Residential Buildings

• HVAC: ASHRAE 62.1, 62.2, and Air Movement

• International Energy Conservation Code

• U.S. Green Building Council LEED v4

• California Building Energy Efficiency Standards: Title 24

Page 56: HVAC for Data Centers

Bill Kosik, PE, CEM, BEMP, LEED AP BD+C,HP Enterprise Business, Technology Services,Chicago, Ill.

Tom R. Squillo, PE, LEED AP, Environmental Systems Design Inc., Chicago, Ill.

Moderator: Jack Smith, Consulting-Specifying Engineer and Pure Power, CFE Media, LLC

Presenters:

Page 57: HVAC for Data Centers

Thanks to Today’s Webcast Sponsors:

Page 58: HVAC for Data Centers

Webcasts and Research

• Modular data center design

• HVAC: ASHRAE 62.1, 62.2, and Air Movement

• 2013 HVAC, BAS state of the industry report

Page 59: HVAC for Data Centers

HVAC for Data Centers

Sponsored by:

Join the discussion about this Webcast on Twitter at #CSEdatacenterHVAC