ccme oct 2015

5
October 2015 44 Keeping IT cool MARKET FEATURE DATA CENTRES As data centres try to keep pace with rapidly changing usage patterns, so do their cooling techniques. Industry experts share insights and updates on best practices and technologies. By Rajiv Pillai | Features Writer T o state the importance of data centres in the present day of quick connectivity and information dissemination is to state the obvious. As data centres try to keep pace with rapidly changing usage patterns, so do their cooling techniques. Pierre Havenga, Managing Director at Emerson Network Power for the Middle East and Africa region, gives an interesting analysis: “If we look at the telecom industry, a few years ago, approximately 90% was voice-centric and 10% was data-centric. Now, it’s approximately 70% data- centric and 30% voice-centric. This is mainly driven by applications on smartphones and people are spending more time downloading apps or text messaging, creating demand for storage of data. You don’t record all the voice communications between people on cell phones, but you need to record all the data. And that’s what’s driving the need for data centres.” In light of this, maintaining and cooling data centres has gained primacy. A whitepaper by Emerson Network Power, a business of Emerson, reveals that cooling systems – comprising cooling and air movement equipment – account for 38% of energy consumption in data centres. 1 As Havenga puts it simply, “You have to reject heat from the data centre; servers generate heat and heat has to be rejected.” Don’t lose your cool Cooling failure is not an option for data centres. In Havenga’s view, the loss of revenue could amount to millions of dollars per day if a data centre is unavailable, with the losses being different for different industries. “For example, for the telecom industry, the losses are quite huge,” he says. However, that could be the least of the problems. As Bart Holsters, A temporary failure of an airport data centre is certainly much more critical than the temporary failure of Twitter’s data centre, though some may argue that

Upload: mohammad-abusaa

Post on 10-Feb-2017

139 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CCME Oct 2015

October 201544

Keeping IT cool

MARKET FEATURE

DATA CENTRES

As data centres try to keep pace with rapidly changing usage patterns, so do their cooling techniques. Industry experts share insights and updates on best practices and technologies.

By Rajiv Pillai | Features Writer

To state the importance of data centres in the present day of quick connectivity and information

dissemination is to state the obvious. As data centres try to keep pace with rapidly changing usage patterns, so do their cooling techniques. Pierre Havenga, Managing Director at Emerson Network Power for the Middle East and Africa region, gives an interesting analysis: “If we look at the telecom industry, a

few years ago, approximately 90% was voice-centric and 10% was data-centric. Now, it’s approximately 70% data-centric and 30% voice-centric. This is mainly driven by applications on smartphones and people are spending more time downloading apps or text messaging, creating demand for storage of data. You don’t record all the voice communications between people on cell phones, but you need to record all the data. And that’s what’s driving

the need for data centres.” In light of this, maintaining and cooling data centres has gained primacy.

A whitepaper by Emerson Network Power, a business of Emerson, reveals that cooling systems – comprising cooling and air movement equipment – account for 38% of energy consumption in data centres.1 As Havenga puts it simply, “You have to reject heat from the data centre; servers generate heat and heat has to be rejected.”

Don’t lose your coolCooling failure is not an option for data centres. In Havenga’s view, the loss of revenue could amount to millions of dollars per day if a data centre is unavailable, with the losses being different for different industries. “For example, for the telecom industry, the losses are quite huge,” he says. However, that could be the least of the problems. As Bart Holsters,

A temporary failure of an airport data centre is certainly much more critical than the temporary failure of Twitter’s data centre, though some may argue that

Page 2: CCME Oct 2015

October 2015 45

Operations Manager at Cofely Besix Facility Management, points out, a cooling failure will result in loss of uptime, with the servers eventually shutting down and the electronic equipment getting damaged.

Mohammad Abusaa, a Business and Project Development Professional with HH Angus and Associates, and a veteran when it comes to data centre cooling, presenting a clear picture of the stakes involved in case of cooling failure in various sectors, says: “The critical nature of cooling for a data centre can be understood from the fact that in many cases, losing cooling for less than five minutes could cause the IT equipment to fail. In some high-density applications, the time could be less than two minutes. The criticality of IT systems’ failure is gauged by the function of the data centre. In other words, a temporary failure of an airport

data centre is certainly much more critical than the temporary failure of Twitter’s data centre, though some may argue that.” The reason for such cooling failure, he says, can directly be related to the cooling system itself, such as the failure of pumps, fans or chillers, and, at times, indirectly related to the cooling system, such as power outages.

Abusaa elaborates that when failure occurs in the cooling system, standby equipment or paths are brought online to ensure continuous supply of cooling to the IT space. Therefore, attributes like redundancy and standby should be factored in at the design stage of cooling systems. When failure occurs in the power supply, an Uninterruptible Power Supply (UPS) device connected to the critical parts of a cooling system – usually the distribution components – will maintain the operation of the cooling distribution network, while the backup generators come online, thereby providing sufficient power to bring the

cooling generation system back online within minutes of losing power. In Abusaa’s view, this is the usual contingency procedure in case of a cooling failure.

Cooling solutionsSince cooling is an imperative for data centres, ASHRAE has defined standards for their cooling requirements, which normally dictates the operating conditions.

Håkan Lenjesson, Market Area Director at Systemair for the Middle East and Turkey region, says that ASHRAE has been broadening the operating ranges and also recommending a very low Power Usage Effectiveness (PUE). Havenga adds: “Today, ASHRAE’s recommended conditions range from 18 degrees C to 27 degrees C. However, the allowable range can even go up to 35-40 degrees C, depending on the server technology.”

With the ranges and requirements defined, the next step is to decide on a cooling solution, as there are several

options available in the market. Havenga says that most data centres currently adopt the traditional direct expansion technology, which is applicable everywhere in the world. “Then there is free cooling, where you have fresh air coming directly from outside,” he points out, and adds, “Or indirect cooling, where you are cooling a medium, typically water. Even further, there are adiabatic solutions, which are an enhancement of the cooling capacity of the chiller. It increases the free cooling capacity.”

Havenga reveals that the latest technology in the market is evaporative free cooling. “This is a way of cooling a data centre without using a compressor; so basically, all-year-round cooling,” he says. He believes that the main driver behind this is energy savings, adding that energy is the single biggest cost incurred by a data centre, which has led to various advancements in technology, InRow cooling being one of

Image so

urce

: Håka

n Le

nje

sson

Bart Holsters

Pierre Havenga

Mohammad Abusaa

Håkan Lenjesson

Figure 1

Page 3: CCME Oct 2015

DATA CENTRES

MARKET FEATURE

October 201546

Håkan Lenjesson says

that most organisations

continue to plan and

design new computing facilities

without much change or

innovation. For example, first,

they design a building and leave

some portion for the data hall

or whitespace. Then, they fill the

whitespace with as many server

racks as possible. In Lenjesson’s

opinion, designing data centres

in the traditional manner can

create a wide range of problems.

He explains, “For example,

an undersized or oversized power and cooling

infrastructure can limit operating capacity or

increase capital expenses.”

He believes that large corporations are looking

for some extra free cooling, while keeping the

PUE as low as possible.

“Companies like Google, Facebook, etc., are

building their new data centres in the very

A walk through Emerson Network

Power’s Customer Experience Centre,

Dubai, will make you realise that

data centre cooling is a serious business.

Climate Control Middle East visited the

facility where Pierre Havenga demonstrated

that cooling in data centres was more

about managing the heat rather than

cooling. He explained: “Five years ago, you

allowed a certain amount of cool air into

your data centre, irrespective of whether

it was required or not. Nowadays, with

Electronically Commutated (EC) fans,

software and wireless monitoring, we can

manage the amount of cool air based on

what is required by the data centre. So, you

don’t need to provide cool air if there is no

heat.

“So our other solution is that you can

even switch off servers. If the server fan is

running, then the fan consumes power, and

the power generates heat. So, it’s getting

to that level of managing your heat levels.

That’s why we call it Thermal Management.

We don’t call it cooling anymore, because

now you manage the thermal side of your

data centre. We manage the temperature

requirement from the rack back to the

chiller.”

Flying north

“We don’t call it cooling anymore”

them. Havenga explains that InRow cooling is a type of air conditioning system commonly used in data centres, in which the cooling unit is placed between the server cabinets in a row for providing cool air to the server equipment more effectively.

Abusaa puts in a nutshell a few of the current trending innovations in the market: “There is modularisation – the main drivers behind this are quality control, cost and delivery schedule. Then there is Direct Liquid Cooling, followed by a continuous development of IT hardware systems that run at higher temperatures and humidity levels and, finally, the on-site Combined Heat and Power (CHP) systems.”

Making the server serveAmidst these cooling options, there lie several challenges, with availability and uptime being the primary ones. “You have to make a redundant solution, no matter where you are in the world,” Havenga says. “This is required so that you will not lose production, because if you lose production, you will lose millions every day.”

Abusaa points out that cooling load or capacity management is one of the most critical challenges in designing a data centre cooling system.

north places of the planet,

for example, to north of

Sweden, Finland, Canada,

etc.,” he reveals. He elaborates:

“Instead of under- or over-

provisioning their new facility’s

power and cooling resources,

companies are installing the

optimal infrastructure for the

precise array of hardware and

enclosures they’ll be using.

Instead of improvising solutions

for efficiency-sapping structural

defects, they’re preventing

those defects from occurring in

the first place. The end-result

is a data centre that’s not only less costly

to cool and maintain but also more reliable

and better suited for business requirements.

They can also allow for significantly increased

intensity usage. ROI is a very important

factor, so doing it right from the beginning is

essential; this is where modular systems are

coming into place.”

A Technical Manager operating on Emerson Network Power’s

Liebert CRV Thermal Management Unit (© 2015 Emerson Electric Co)

Emerson Network Power’s Customer Experience Centre

For security and access reasons, the CRAC unit might need to be relocated

Page 4: CCME Oct 2015

October 2015 47 www.dencohappel.com

GEA Middle East LLC

Technopark Head Quarters Building,

Block B, 3rd Floor, Office 321,

Technopark, P.O. Box 261236,

Dubai, United Arab Emirates

Scott Ross

Director of Sales

MENA & APAC

Mobile: +971 50 5548502

Office: +971 4 881 6065

Air Treatment Holding

becomes

Page 5: CCME Oct 2015

DATA CENTRES

MARKET FEATURE

October 201548

“Hot Huts”

Google has designed custom cooling systems for

its server racks. The systems are called “Hot Huts”,

because they serve as temporary homes for the hot

air that leaves the servers – sealing it away from the

rest of the data centre floor. Fans on top of each Hot

Hut unit pull hot air from behind the servers through

water-cooled coils. The chilled air leaving the Hot Hut

returns to the ambient air in the data centre, where

the servers can draw the chilled air in, cool them down

and complete the cycle.

Evaporative cooling

As hot water from the data centre flows down the

towers through a material that speeds evaporation,

some of the water turns to vapour. A fan lifts this

vapour, removing the excess heat in the process, and

the tower sends the cooled water back into the data

centre.

Using seawater

Google’s facility in Hamina, Finland, uses seawater

to cool without chillers. The company has chosen

Hamina for its cold climate and its location on the

Gulf of Finland. The cooling system pumps cold water

from the sea to the facility, transfers heat from the

operations to the seawater through a heat exchanger,

and, then, cools this water before returning it to the

Gulf. Since this approach provides all the needed

cooling year round, Google claims to not have installed

any mechanical chillers.

How Google does it

He stresses that capacity management is related to the phasing of data centre construction, expansion or phasing out of IT loads within the data centre, variation in the cooling load profile within the day/month/year and variation in the cooling load requirements within the same data hall or even at the server rack level.

The other challenges are maintenance-related. Abusaa stresses that it is crucial to understand that a facility that is running 24/7 and has strict security access regulations will have its design and operation challenges when it comes to maintenance. He says, “For example, while having a Computer Room Air Conditioning Unit (CRAC) installed in a specific location is the most efficient solution, for security and access reasons, the CRAC unit might need to be relocated to ensure that the maintenance personnel do not have access to the IT equipment, as there might be a possibility of accidently damaging the IT equipment while maintaining the CRAC unit.”

And then, there is always the issue related to humidity (See Figure 1) and air quality. “Similar to hospitals and other

At Google data centres, the company often uses water instead of chillers, as an energy-efficient way

to cool…

(Information source: https://www.google.ae/about/datacenters/efficiency/internal/#water-and-cooling)

“IT, data centres and server

equipment consume electricity

and emit heat as a ‘waste

product’,” says Ziad Youssef, Vice

President of IT Business - UAE, Gulf

Countries at Schneider Electric. “At such

an enclosed environment with sensitive

technology, the heat can be damaging.”

As data centres experience exponential growth in the region,

he believes that new solutions to curtail the simultaneous rise

in energy costs will be essential. “One such solution is cooling

– which is critical to the smooth functioning of a data centre,

and to the maintenance of hardware carrying mission-critical

enterprise data,” he says.

Combating ‘waste’ heat

Ziad Youssef

Similar to hospitals and other critical facilities, maintaining control of the Indoor Air Quality to avoid contamination is crucial

critical facilities, maintaining control of the Indoor Air Quality to avoid contamination is crucial,” Abusaa reveals, and adds, “This is not only achieved through filtration but also through the design of data centres and operation and maintenance guidelines.” He is emphatic that filtration, dehumidification, access control and other practices should address the air-quality issue.

Reference1. http://www.emersonnetworkpower.com/documentation/en-us/brands/liebert/documents/white%20papers/enterprise-data-center_24622.pdf