Featured Video, News

Microsoft to build duo of UAE data centres

0 17Microsoft has announced plans to build its first Middle East and North Africa data centres in Dubai and Abu Dhabi.
Don't show me again

Data center that bets on 100% uptime

A few miles from the glitzy casinos of the Las Vegas strip stands a highly secure, 407,000-square foot building which — according to the man who operates it — is the most energy efficient, high-density data center in the world.

Rob Roy, the CEO, founder and chairman of Switch Communications Group, is walking the halls of his seventh and most impressive data center, the SuperNAP, from which he provides co-location services to some of the world's biggest organizations. At 3 p.m., one hour into giving a tour, it's clear that Roy is not a man who easily runs out of energy.

“Do we have a time limit?” he asks. “This is my last thing of the day, so I'll just talk till midnight.”

The tour ends by 4 p.m., eight hours early, but Roy has plenty of time to explain why the SuperNAP is a safe bet for organizations with the strictest uptime and security requirements. What makes the SuperNAP so interesting? Here are some of the highlights:

Guaranteed 100% uptime

Five nines of availability doesn't impress Roy. “We give 100% service-level agreements, guaranteed,” he says. “Obviously, that's a big monetary risk if I didn't feel this design was ready for that. Our NAP 4 facility, which is our next biggest site [and also in Las Vegas], for three years has had 100% uptime.”

The SuperNAP (network access point) operates its own 250 megavolts ampere (MVA) substation, 146 MVA of generator capacity and 84 MVA of UPS (uninterruptable power supply), topped off with 30,000 tons of redundant cooling.

“Our network for six years has never had an outage,” Roy says. “Every single part of it fails. In any given month, something fails. Blades fail, Cisco routers fail, carriers fail, Sprint fails, Verizon fails, AT&T fails. But we build this stuff in such a redundant manner. The chance of something happening with us where you have an outage is really less than anywhere from a design standpoint.”

Militant security

Switch began life in 2000 as a government contractor and has built seven data centers meeting the high levels of security demanded by government and military clients. Military-trained security staff protects the SuperNAP, with at least three guards on site at any given moment.

“By the time you got into this building, you went through our $2 million blast wall, live armed security and six or seven layers,” Roy says. “You went through biometrics.” Although customer equipment is locked inside cages, Roy says the cages are superfluous. “I really believe at this site we would never have to put a cage on anything because by the time you're here you've been so researched.”

Switch has built all seven of its data centers in Vegas because of the region's relative lack of natural disasters. The SuperNAP is six miles from the Vegas airport, far enough to stay out of the path of incoming and outgoing planes, but near enough to be in a designated no-flight zone, guaranteeing that no planes will fly over the data center, according to Roy.

One other nice touch helps keep the SuperNAP equipment free of dirt and particles: Visitors must frequently walk over sticky white mats, which grab dirt from the bottoms of one's shoes and are common in the military, aerospace, microelectronics, pharmaceutical and hospital industries.

Focus on power and cooling

How to power and cool a data center efficiently is top of mind for most people in the industry today. Roy, as is his wont, claims to do so more efficiently than anyone else.

Data-center efficiency is evaluated with PUE, or power usage effectiveness, a metric devised by the Green Grid which essentially compares the total power needed to run a facility with the power needed to run data-center equipment.

According to a Google research paper on energy efficiency, “a PUE of 2.0 indicates that for every watt of IT power, an additional watt is consumed to cool and distribute power to the IT equipment.”

A typical data center's rating is above 2.0, says Daniel Tautges, U.S. president of data-center management vendor GDCM.

Google has claimed an average PUE of 1.21 across six data centers, and the most efficient of the six had an annual rating of 1.15 and a best quarterly rating of 1.13. These numbers are “fantastic,” Tautges says. If the numbers are accurate, only about two-tenths of a watt (for every watt of IT power) is needed to cool Google's servers and distribute power to IT equipment.

Roy claims the SuperNAP's average PUE rating is 1.146, just a hair better than Google's best annual rating. He also says he expects the Green Grid to certify his contention that the SuperNAP is the world's most efficient, highest-density data center.

Numbers aside, Roy lists many factors that make the SuperNAP unusually efficient and dense. Instead of using indoor computer room air conditioner (CRAC) units, the SuperNAP has 600-ton air handlers plugged into the SuperNAP from the outside. Specially designed software constantly analyzes the environment and automatically switches between four types of chiller systems depending on the time of day, temperature and moisture. DX cooling, chilled water, indirect evaporative and direct evaporative cooling are all in use.

“That unit adjusts moisture in the building many times faster than any CRAC unit ever could,” Roy says. “We have almost 6 million cubic feet per minute of air we can push [through the data center]. We can change the air in this entire data center every two minutes.”

You might be wondering why a building with such massive cooling needs would be located in the Las Vegas desert. But dry air is the most efficient for cooling, and the temperature drops at night and during the winter, Roy says, adding that the temperature is below 68 degrees Fahrenheit in the Vegas Valley for more than half the year (5,000 hours).

Because the cooling is so efficient, Roy can deliver watts to servers at densities of 1,500 watts per square foot. Typical data centers built today can handle heat loads of only 350 to 500 watts per square foot, according to AMD.

According to Roy, customers at other co-location centers might be limited to 250 watts per square foot, an amount that can be exceeded with blade servers and other modern IT products.

If the SuperNAP is delivering far more power per square foot than other data centers, wouldn't that negatively impact the facility's efficiency? Roy says no. “Efficiency has nothing to do with consumption,” he says. “You can consume a lot efficiently or inefficiently, or you can consume the most in the world efficiently or inefficiently.”

Previous ArticleNext Article

Leave a Reply

Your email address will not be published. Required fields are marked *


The free newsletter covering the top industry headlines