The Brain Of The Beast: Google Reveals The Computers Behind The Cloud
Behind the ephemeral "cloud" of cloud computing, the network we use for everything from checking our email to streamlining our health care system, there lies a very tangible and very big computer infrastructure.
But besides a glimpse at some of the hardware in 2009, there has been little information about Google's data centers, the warehoused collections of servers that have given the company the foundation for its vast Internet operations.
Today, the company is throwing open the gates to the world — digitally, of course. It has released a site featuring photos of facilities from Belgium to Finland to Iowa and launched a guided Street View tour of one in Lenoir, N.C.
It's the same facility the company revealed in an exclusive tour to Wired senior writer Steven Levy, whose story on Google's infrastructure appears in the magazine's November issue. In an interview with Morning Edition's Steve Inskeep, Levy described going where no Google outsider has gone before:
"What strikes you immediately is the scale of things. The room is so huge you can almost see the curvature of Earth on the end. And wall to wall are racks and racks and racks of servers with blinking blue lights and each one is many, many times more powerful and with more capacity than my laptop. And you're in the throbbing heart of the Internet. You really feel it."
Google has a lot of servers — Levy reports 49,923 in Lenoir alone. The total worldwide number eludes even Google's executives, but there have been at least 1 million cumulatively, according to a plaque Levy found on the premises.
Google is not the only company with massive computer networks. In 2006, Microsoft built a giant data center on 75 acres of bean fields in Quincy, Wash. Yahoo and Dell, among other companies, have also set up data centers in Quincy.
But Google's data center technology is unique, Levy says, which is partly responsible for the company's success — and a cause for the secrecy surrounding it.
"One technique that Google really pioneered was keeping things hotter than has been traditionally expected in a data center," Levy says. "In old data centers, you would put on a sweater before you went in there. Google felt that you could run the general facility somewhat warmer than even normal room temperature. When I walked into Lenoir I think it was 77 degrees."
The trick to keeping the heat under control, Levy writes, is a "hot aisle," a space that harnesses hot air from the servers into water-filled coils, sends it out of the building to cool, and then circulates it back inside. This is a dramatic departure from traditional centers, which use large amounts of energy on air conditioning.
Levy says that the specific technology varies from center to center and that Google has considered the local resources and geography in each design. That doesn't necessarily mean, however, that the centers don't have an environmental cost.
"There's no way around it. These things burn a lot of energy, and a lot of the energy in a data center is done to cool it down so the computers don't melt. Data centers in general consume 1.5 percent roughly of all the world's electricity," Levy says.
Google's servers have been getting progressively faster and cheaper, and even now the company has plans to completely change the basics of the system in places like Lenoir. But the specifics of those changes will remain secret — at least for now.
"Google may be dedicated to providing access to all the world's data, but some information it's still keeping to itself," Levy writes.
STEVE INSKEEP, HOST:
Next, we'll visit the Internet cloud. That's a term people often use to describe the place or places where we store stuff online. Increasingly, our databases, email accounts, other information are not stored in local computers, but in giant low-rise warehouses packed with computer servers. They're all over the country, all around the world and they are huge consumers of energy.
Google has around 20 of these data centers, and recently allowed technology writer Steven Levy into one of them in North Carolina to show off how energy-efficient Google is trying to be. It gave Levy a glimpse into the online world as people rarely get to see it. And he writes about his experience in the new issue of Wired magazine.
STEVEN LEVY: What strikes you immediately is the scale of things. The room is so huge you can almost see the curvature of Earth in the end. And it's wall to wall, are racks and racks and racks of servers with blinking blue lights and each one is many, many times more powerful and with more capacity than my laptop. And you're in the throbbing heart of the Internet. And you really feel it.
INSKEEP: So you're in this data center. It's using energy quite efficiently compared to the average data center, even a pretty good one. What are some of the techniques that are used and what do they look like?
LEVY: Well, one technique that Google really pioneered was, you know, keeping things hotter than has been traditionally expected in a data center. In old data centers, you would put on a sweater before you went in there. Google felt that you could run the general facility somewhat warmer than even normal room temperature. When I walked into Lenoir, I think it was 77 degrees.
INSKEEP: And that doesn't run down the computing equipment?
LEVY: Computer equipment is actually tougher than people expect. And they isolate the really hot air that comes out from the back of the servers into what's known as a hot aisle, and that's sealed off and it's maybe 120 degrees, and that's where they take that very hot air and do the water cooling.
Google takes a look at the geography and the resources every time they build a data center. So in North Carolina, they did something that was relatively traditional. They have these coolers where the water circulating goes outside and cools down before it reenters the data center to cool down the servers. But in Finland, which I did visit, they use seawater to cool the data center.
INSKEEP: Now this is a huge issue because computers generate so much heat that keeping it cool would be a tremendous use of energy, a tremendous waste of energy in the view of some.
LEVY: There's no way around it. These things burn a lot of energy, and a lot of the energy in a data center is done to cool it down so the computers don't melt. Data centers in general consume 1.5 percent, roughly, of all the world's electricity.
INSKEEP: So as you're talking, I'm thinking about cloud computing, the Internet cloud. And many of us are getting used to this idea that if we have an email account, it might not be saved in the machine where we are at; it's going off somewhere. But once you actually got in to look at one of these places and hear it, feel it, did it change your perceptions of what's going on in the world when you went back to your computer screen at home?
LEVY: It actually did. You know, many, many years ago I went on a journalistic quest for Einstein's brain, which was lost then. And I felt if I saw it, it might be an anticlimax. But when I actually did see it, it really opened up my eyes; it was a revelation. This is where, you know, the power of the atom came from and relativity and all those other things. And I had the same kind of experience inside that Google data center. Here was the ephemeral made real, you know, the cloud really was something and it was something quite remarkable and breathtaking.
INSKEEP: Steven Levy, thanks very much.
LEVY: Thank you.
INSKEEP: He's a senior writer for Wired magazine. Transcript provided by NPR, Copyright NPR.