News

'68-degree data centers becoming a thing of the past'

Cooling a data center to 68 degrees may be going out of style, APC power and cooling expert Jim Simonelli says.

Servers, storage and networking gear are often certified to run in temperatures exceeding 100 degrees, and with that in mind many IT pros are becoming less stringent in setting temperature limits.

Servers and other equipment “can run much hotter than people allow,” Simonelli, the chief technical officer at the Schneider Electric-owned APC, said in a recent interview. “Many big data center operators are experienced with running data centers at close to 90 degrees [and with more humidity than is typically allowed]. That’s a big difference from 68.”

Simonelli's point isn’t exactly new. Google, which runs some of the country’s largest data centers, published research two years ago that found temperatures exceeding 100 degrees may not harm disk drives.

But new economic pressures are helping data center professionals realize the benefits of turning up the thermostat, Simonelli says. People are starting to realize they could save up to 50% of their energy budget just by changing the set point from 68 to 80 degrees, he says.

Going forward, “I think the words 'precision cooling' are going to take on a different meaning,” Simonelli says. “You’re going to see hotter data centers than you’ve ever seen before. You’re going to see more humid data centers than you’ve ever seen before.”

With technologies such as virtualization increasingly placing redundancy into the software layer, the notion of hardware resiliency is starting to become less relevant, reducing the risk of over-heating.

Server virtualization also imposes new power and cooling challenges, however, because hypervisors allow each server to utilize much greater percentages of CPU capacity. On one hand, server virtualization lets IT shops consolidate onto fewer servers, but the remaining machines end up doing more work and need a greater amount of cold air delivered to a smaller physical area.

If you're shutting off lots of servers, a data center has to be reconfigured to prevent cooling from being directed to empty space, Simonelli notes.

“The need to consider power and cooling alongside virtualization is becoming more and more important,” he says. “If you just virtualize, but don’t alter your infrastructure, you tend to be less efficient than you could be.”

Enterprises need monitoring tools to understand how power needs change as virtual servers move from one physical host to another. Before virtualization, a critical application might sit on a certain server in a certain rack, with two dedicated power feeds, Simonelli notes. With live migration tools, a VM could move from a server with fully redundant power and cooling supplies to a server with something less than that, so visibility into power and cooling is more important than ever.

The ability to move virtual machines at will means “that technology is becoming disconnected from where you have appropriate power and cooling capacity,” Simonelli says.

To support the high densities introduced by virtualization and other technologies such as blade servers, cooling must be brought close to the rack and server, Simonelli says. As it stands, cooling is already the biggest energy hog in the data center, with power wasted because of over-sized AC systems and temperatures set too low, he says.

While every data center has different needs, Simonelli says enterprises can learn something from the giant SuperNAP co-location data center in Las Vegas, a 407,000 square-foot building that relies heavily on APC equipment, such as NetShelter SX racks, thousands of rack-mounted power distribution units and UPS power supplies.

While the site can support 60 megawatts, it’s being built out in 20-megawatt chunks. “That means they can maximize the energy consumption, the efficiency of the data center as they scale. They’re not powering the 60-megawatt site right away,” Simonelli says.

One of the biggest mistakes is to over-size power capacity, in anticipation of future growth that may never come. Companies have to plan for where they think they will be a few years from now, but build out in smaller increments, he says.

“You have to have the floor space, and you have to have the capability of getting power from the utility,” Simonelli says. “But if you’re going to build out a one- or a five-megawatt data center, and you know that your first year of deployment is only going to be 100 kilowatts, get the space and make sure you have power from the utility for five megawatts but just build it out in 250 or 500 kilowatt chunks.”

Previous ArticleNext Article

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

GET TAHAWULTECH.COM IN YOUR INBOX

The free newsletter covering the top industry headlines