Green computing is a hot-button issue right now, but not all the ideas out there are practical for data centers. “It’s 90 percent hype,” says Ben Stewart, senior vice president of facilities planning at Terremark Worldwide Inc. He’s dubious about solar and wind power, for example. But Stewart says 10 per cent of the ideas are win-win: Done right, certain green initiatives can increase energy efficiency, reduce carbon emissions and yield savings.
According to Steve Sams, vice-president at IBM Global Technology Services, there’s only one way to evaluate green energy options. “If I spent the money, where would I get the best return? That’s the question to ask,” says Sams. The key is knowing where to start. These four questions and answers can help you develop a plan.
Q: Why should I care about having a green data centre?
A: Data centre managers who have run out of power, cooling or space are already motivated to move to greener practices. But many others don’t care because they put reliability and performance first — and they don’t see the power bills, says Peter Gross, CEO at New York-based EYP Mission Critical Facilities Inc. That’s likely to change as electricity consumption continues to rise. “Our data centres are a small fraction of our square footage but a huge percentage of our total energy bill,” says Sams.
The cost of electricity over a three-year period now exceeds the acquisition cost of most servers, says Gross. “I don’t know how anybody can ignore such an enormous cost. It is the second-largest operating cost in data centres after labor,” he says. Gross says that every CIO, facility manager and CEO he meets expresses concern about data center energy efficiency.
“My CEO is beating the drum about cutting power consumption,” says John Engates, chief technology officer at hosting company Rackspace Inc. in San Antonio. He says just 50 per cent of power coming into the data centre goes to the IT load. The rest is consumed by surrounding infrastructure, including power, cooling and lighting. “If you’re using less power, you’re spending less money. It’s just good business,” Engates says.
Returns on investment can be difficult to determine, however, because in most cases, the IT staff in a data center doesn’t see the power bill. “The single most important step is to find ways to measure efficiency in your facility,” says Gross. “You cannot control what you cannot measure.” One way to determine overall data center energy efficiency and provide a benchmark is to hire professionals to do an analysis. An inspection by IBM Global Technology Services costs US$50,000 to $70,000 for a 30,000-square-foot data centre, says Sams.
But just a one- or two-day engagement might get you most of the benefits for a lot less money, says Rakesh Kumar, an analyst at Gartner Inc. “You can get 80 per cent accuracy with a small investment in consultancy costs,” he says. “That’s good enough to make some judgments.”
Q: What steps can I take to increase the efficiency of my data center’s IT equipment?
A: The biggest savings come from server consolidation using virtualization technology. Not only does this remove equipment from service, but it also helps raise server utilization rates from the typical 10 percent to 15 percent load today, increasing energy efficiency.
Consolidating onto new servers brings an additional benefit. Power-supply efficiencies for servers purchased more than 12 months ago typically range from 55 per cent to 85 per cent, says Gross. That means 15 per cent to 45 per cent of incoming power is wasted before it hits the IT load. Newer servers operate at 92 per cent or 93 per cent efficiency, and most don’t drop below 80 per cent, even at lower utilization levels.
Using virtualization, Affordable Internet Services Online Inc. in Romoland, Calif., consolidated 120 servers onto four IBM xSeries servers. “Now we don’t have the power use and cooling needs we had before,” says CTO and co-founder Phil Nail.
Using networked storage can also keep energy costs in check. Direct-attached storage devices use 10 to 13 watts per disk. In an IBM BladeCenter, for example, 56 blades can use 112 disk drives that consume about 1.2 kilowatts of power. Those can be replaced with a single 12-disk Serial Attached SCSI storage array that uses less than 300 watts, says Scott Tease, BladeCenter product manager.
IT managers should demand more energy-efficient designs for all data centre equipment, says Engates. He says his company standardized on Brocade Communications Systems Inc. switches in part because of their energy efficiency and “environmental friendliness.”
Q: How can I get more out of my data centre’s cooling and mechanical systems?
A: Getting back to basics is key, says Dave Kelley, manager of application engineering at Columbus, Ohio-based Liebert Precision Cooling, a division of Emerson Network Power Co. “You have to go back and look at a lot of the things that you didn’t worry about 10 years ago.”
The biggest potential savings come from airflow optimization. For every kilowatt of load, each rack in a data center requires 100 to 125 cubic feet of cool air per minute. Airflow blockages under the floor or air leaks in the racks can cause substantial losses, says Kelley. The typical response to such problems has been to increase the air conditioning temperature — and that’s a big energy-waster.
Simple steps such as implementing hot-aisle/cold-aisle designs, sealing off cable cutouts, inserting blanking plates and clearing underfloor obstructions make a big difference. With greater airflow efficiency, air conditioning output temperatures can be raised.
After performing a computerized airflow analysis of its data centers, San Francisco-based Wells Fargo & Co. did exactly that. “In many data centers, you can hang meat in there, they’re so cold. With computerized control and better humidification systems, we’ve raised the set point of our data centers so we’re not overcooling them,” says Bob Culver, senior vice president and manager of facilities for Wells Fargo’s technology information group.
At Pacific Gas and Electric Co. (PG&E), cable races under the floor were blocking 80 percent of the airflow. The utility expects to save 15 percent to 20 percent in energy costs by rewiring under the floor, redesigning the return-air plenum and carefully choosing and placing perforated tiles in the cold aisles. Choosing the right perforated tile — a seemingly small consideration — can actually make a big difference. “There are better tiles out there that will give you more efficient distribution of cool air,” says Jose Argenal, PG&E’s data centre manager. The changes also allowed PG&E to avoid adding chillers, pumps and piping — and piping is a potential problem in its older, basement-level data center.
Data centre managers can also optimize air conditioning systems by using variable-speed fans, says Ken Baker, data center infrastructure technologist at Hewlett-Packard Co. “AC runs at 100 percent duty cycle all the time, and the fans have one speed: on,” he says. HP’s Dynamic Smart Cooling initiative uses rack-mounted temperature sensors, and variable-speed fans allow the power consumption of air conditioning units to vary with the IT equipment load. Intelligent control circuitry manages both fan speed and temperature settings on air conditoners.
It’s relatively easy to retrofit existing fans, Baker says, and the approach has two major benefits. One is that cutting fan speed dramatically reduces energy use. A 10-horsepower fan uses 7,500 watts of power at full speed but just 1,000 watts at half speed, he says. The increased efficiency also allows the temperature of the cool air supply to be automatically raised from the typical 55 degrees Fahrenheit to between 68 and 70 degrees, he says.
“The biggest low-hanging fruit is just turning the thermo