Building the Green Data Center 86
blackbearnh writes "O'Reilly News talked to Bill Coleman, former founder of BEA and current founder and CEO of Cassett Corporation, about the challenges involved in building more energy-efficient data centers. Coleman's company is trying to change the way resources in the data center are used, by more efficiently leveraging virtualization to utilize servers to a higher degree. In the interview, Coleman touches on this topic, but spends most of his time discussing how modern data centers grossly overcool and overdeploy hardware, leading to abysmal levels of efficiency."
Lean Code = Green Code (Score:5, Interesting)
Software has an impact, too. Messy, heavy code takes longer to run, takes more CPUs, etc. Imagine how much energy could be saved if there wasn't so much code bloat!
Re: (Score:3, Insightful)
Software has an impact, too. Messy, heavy code takes longer to run, takes more CPUs, etc. Imagine how much energy could be saved if there wasn't so much code bloat!
So that means that servers should be built the Gentoo way, from scratch, using just the things you need, no more, no less.
How much does it cost to deploy such a server?
How much does it cost to pay someone qualified enough to do it properly?
The code bloat is paired with feature bloat. And the more features there are, the more you have to pick and choose -- or, if you cannot choose, support. Because your users will want them, more likely than not.
Now, cleaning up the world's code... sounds like great wo
Re: (Score:2)
o that means that servers should be built the Gentoo way, from scratch, using just the things you need, no more, no less.
How much does it cost to deploy such a server?
How much does it cost to pay someone qualified enough to do it properly?
Frankly anyone with half a brain can pretty much use mkinitrd to make such a server.
How much does it cost to hire 500 admins for thousands of machines rather than half a dozen? How much does electricity and AC cost?
Meh, no point explaining. The price of oil and the economics will do that job.
Re: (Score:2)
possibly not, perhaps he means that software should be built without the 'make it easy for the developer' features that modern languages contain. I mean, its easy to write an app in a scripting language but it will be bigger, slower, require a VM to host the script, will use more memory (especially so if it has a garbage collector), and so on.
There is a trend of saying that programmer productivity is everything, and if it requires faster computers with more RAM, then that's just too bad. I'm sure that one d
Re: (Score:3, Funny)
Messy, heavy code takes longer to run, takes more CPUs, etc.
Do you guys have to bring Vista into every thread?
Re: (Score:1)
YES! ...Duh.
The outback (Score:1, Interesting)
Re: (Score:3, Insightful)
No one in the server farm business is going to try and break into the solar-power business. It's not their area of expertise. It's an entirely different sort of business altogether. If there were a ton of solar power stations littering the outback, or if someone enterprising were ready to put some up in the hopes of attracting power-hungry industries with cheap electricity, that'd be another thing. But I would imagine it's still a rather risky proposition, as far as things go.
Besides, the bandwidth and l
Re: (Score:1)
No one in the server farm business is going to try and break into the solar-power business.
Kinda like how the roof of Google's headquarters isn't covered in solar panels [google.com]?
Maybe they don't want to get into the business of supplying solar panel, but there's plenty of interest in using renewable technology to try and lower energy costs. (Not to mention even photovoltaic solar can help reduce cooling bills, because insolation is being used to generate electricity instead of simply heating the interior.)
Re: (Score:1)
dust is the problem. ever try diagnosing a fault in your operating system when everything is correct, the power supply is operational (and come to find out) that a fine layer of micron-sized dust on the contacts of your graphics card has been adjusting the signal just enough to cause the driver to crash? now multiply that by a thousand machines, and you'll know why nobody is rushing out into the desert to build data centers.
Re: (Score:1)
Re: (Score:2)
I would guess a combination of difficulty and expense in
a) bandwidth
b) cooling
c) supplies
d) available utilities (i.e. running water, available healthcare)
He's missing real world experience (Score:5, Informative)
Re: (Score:3, Interesting)
Of course with a good dynamic provisioning system a single host failure doesn't matter because that new VM will just get spun up on a different host that's woken up.
Bingo. A node is just a node. A decent control system will detect a node failing to come up, flag it for service and bring up another one. In some datacenters not designed for this sort of redundancy a server failure is a big deal where people have to come in on a holiday weekend. If you do it right the dead server just sits there until you get around to that rack in the regular course.
Re: (Score:2)
Or as I've heard an opinion about the Sun's "container server farm", only 90% of the servers are working, and every one that breaks is replaced by one unused before - and when there are not enough spares, the entire container is replaced at the customer
Re: (Score:1)
Re: (Score:2)
Re: (Score:3, Informative)
This "grossly overcooling" business is done for several reasons:
There are 18 Celsius just out from the cooling units, but there might be pockets of warmer air in the data warehouse (based on rack position and use).
This "grossly overcooling" allows the servers to have a long duration of functionality when the air conditioning breaks.
The PSUs are working better at lower temperatures (even if they are perfectly fine otherwise). Also, the cooling
Re: (Score:2)
You might as well say... (Score:2)
"We're a Windows shop." You only hint at your real concerns -- that license tracking and organizational inertia prevents it in your case. That's too bad for you.
The technology is obviously available and immensely powerful. Some will use it, some will shun it. In the corporate world which do you suppose is going to out-compete the other?
Re: (Score:2)
Re: (Score:2)
How do you sell it to the business if their budget is getting squeezed (I know this applies to physical servers as well, but if you have capacity in the virtual cluster to fit their app in it's a lot harder to say no).
Oh, you're looking at it from a salesman's point of view, rather than a customer's. That can't be good for your customer. Since Xen is an open source project RedHat's new approach using KVM [cnet.com] could prove more interesting.
1 dual processor/8 core server running Oracle with in-memory cache option and support: roughly $200,000.
50 dual processor/8 core servers each running several VM's of postgresql with pgpool-II [mricon.com] and memcached [danga.com]: roughly $200,000. The freedom to PXEBoot a blank box into a replicant node faste
Hm, good summary (Score:3, Funny)
I should have printed a fresh stack of these. [bullshitbingo.net]
Managed power distribution units (Score:4, Interesting)
Switch the machines off at the the socket. You can do it using SNMP.
Monitor the average load on your machines, if too low, migrate everything off it and switch a machine off. If too high, switch one on.
Course it assumes you know how to create highly available load balanced clusters. Automatic installations, network booting and all that. Not so difficult.
Re: (Score:2)
This is actually getting remarkably easy for Linux clusters, and the help is coming from a bizarre source - LTSP.
I wrote a journal piece [slashdot.org] about it just recently. I'm setting this up for me and it's interesting.
People are doing some interesting stuff with LTSP -- call centers with IP softphones, render farms. Soon we may see entire infrastructure with redundant servers powering on to serve demand spikes and shutting off when not in use.
Diskless servers (Score:2)
We do something like this, but from scratch rather than using LTSP. It's really not difficult, just a slightly different way of looking at how an operating system and server application should work. Think botnet. It's a fundamental shift in the mathematics of computing infrastructure, from linear or worse to logarithmic.
Re: (Score:2)
I'm liking the LTSP model because I can do it without investing my time writing code. I intend to pull up an on demand render farm without writing a single line of code.
I am also interested in the potential of exploiting the unused resources of desktop computers to turn an entire organization into an on-demand compute cluster and/or distributed redundant storage. Joe the typist doesn't need a quad-core 4GB machine to draft a letter, but as long as he's got one we may as well do something useful under the
Re: (Score:2)
Count me in too. I want config details as well
Re: (Score:2)
Depending on the size of your organization, and how "corporate" the network is (are workstations in software lock-down?), you may have to spend a lot of time designing software that can ensure some level of code authenticity for any deployed work. Otherwise, don't expect to get approval to run this on "most every" random workstation. I had a project where machines would automatically download code and data sets to run as requested, synchronized by a central server. Three things stood in my way: (a) limit
Re: (Score:2)
I always think when I see things like this, we are moving more and more to something like the ideas of
Linux NOW [9fans.net].
Re: (Score:2)
Tiny base OS (linux), booted from PXE/TFTP server running from ramdisk. networking, storage, snmp, ssh, grid engine, "botnet" client and bugger all else. On top of that you run a VM host, Xen, VMware, vserver or whatever fits your requirements. This is the infrastructure platform. It can be rolled out to anything which supports PXE. Tens of thousands of machines if required. They can be functional literally as fast as machines can be fitted into racks.
Basically you don't touch a machine till it comes up and
Re: (Score:1)
Thanks. That's the ticket. I assume if it takes 15-30 minutes to configure, you are downloading and chain booting a disk image. I suppose if I take that route I can preload the disk images on a spare server with one boot image that then puts the server back to sleep. Then when the load comes up the provisioned server can be awakened in short order.
It takes under a minute to bring up my clients because everything runs in the ramdisk so far.
I'd let the load get much lower -- maybe .5 on each cpu before
Re: (Score:2)
Thanks. That's the ticket. I assume if it takes 15-30 minutes to configure, you are downloading and chain booting a disk image
Kind of. The base OS boots and runs 100% from ramdisk, takes about 15 seconds to download and maybe another 30 to boot. About 100Mb or so uncompressed. Doing it from scratch keeps the underlying infrastructure OS small. What takes the time is the application image. It usually has some data attached with which it is packaged, anything from a few hundred mb to gigabytes. The local storage is purely for the application packages or VM images while they're hosted on that machine.
What's got me curious is how to make the management piece redundant and load balanced as well. I'll just have to work on it.
Not sure which load balancing
Re: (Score:2)
What you say is shutting off computers, and then starting them with network starts. My own PSU (450W unit) uses more than 15W when connected to mains - so I switch it off from the mains when not in use.
Now, 15W loss in a 500W PSU when off is a drop in a bucket - yet it might help a bit
Re: (Score:2)
Mouse, monitor, and I think this would be all - I've measured the power on the socket, not specifically the PSU
Re: (Score:2)
The monitor is probably the culprit, not the computer.
Re: (Score:2)
Am I missing something? Outside of a few very large organizations, isnt the operation of a data center separated from the equipment inside the data centers? Dont most data centers rent out space for customers? If this is the case, outside of increasing customers bills for energy consumption, there is nothing that a data center can do to change the way the customer does business. Not every customer is going to find it practical to have a managed virtual server environment, or be ok with allowing systems to b
Re: (Score:2)
Depends on the application. Not everything fits in the LAMP, web server model.
Overcooling? (Score:1, Interesting)
I think this guy confuses heat and temperature. In datacenters, cooling costs are mostly proportional to the heat produced, and have little to do with the temperature you maintain in the steady state.
Re: (Score:3, Informative)
And I think you don't understand thermodynamics either. Cooling to say 18 Celsius when you can happily get away with 25 Celsius will have a big impact on your cooling bill even through you are getting rid of the same amount of heat.
Re: (Score:2)
This cooling thing is a thermal machine - and the best efficiency depends on both the current temperature (inside or outside) and the temperature difference. That is, if you have outside 25 Celsius, you need twice the power to cool at 15 Celsius as opposed to cooling at 20 Celsius
Why cool systems? (Score:2, Interesting)
Re: (Score:1)
Re: (Score:2)
Four steps to a green data center (Score:5, Funny)
2. Paint it green
3. ???
4. Cthulhu
former founder (Score:3, Insightful)
How can you be a former founder of something? Someone else can't come along later and found it again can they?
Re: (Score:1)
Re: (Score:1)
lead worded oddly (Score:2)
He's no longer the founder of BEA? Who is then?
Re: (Score:2)
Yea, but the downside is that you've got to run OSX
Don't get me wrong. OSX is a great operating system for a user. It's probably the best laptop OS in existence. I'm writing this comment on a Powerbook G4 right now, actually. But in the server room, OSX sucks if it has to interact with any non-OSX services.
The very fact that they took things in Unix that had worked for 20 years and broke them for no good reason except they didn't fit their idea of how something should work is asinine.
Granted, the recent rel
Re: (Score:2)
Hey... It's been working for Microsoft.
Dell's tool (Score:2)
I've used Dell's Greenprint Calculator [dell.com] to determine usage in my racks pretty often.
It's got a nice interface and gives you all the energy information you need on their equipment, plus allows you to insert your own equipment's energy profile to calculate total usage.
It's very handy
Re: (Score:2)
Yes this is a very good tool, I tried it when upgrading a couple of servers, and was amazed how much heat output memory modules has.
Good info sources on Green Data Centers (Score:4, Informative)
The Green Data Center Blog [greenm3.com]
Data Center Knowledge [datacenterknowledge.com]
Groves Green IT [typepad.com]
The Big List of Green Technology Blogs [datacenterknowledge.com]
Northern Climates? (Score:3, Interesting)
What I've always wondered is why we don't build more datacenters in colder climates here in north america. Why put huge commercial datacenters in places like Dallas or San Diego (there are plenty in each) when you could place them in Canada or Alaska? In a cold enough climate, you could just about heatsink the racks to the outside ambient temperature and have little left to do for cooling. I suppose the downside is 20ms of extra latency to some places, and perhaps having to put more fiber and power infrastructure in a remote place. But surely in the long run the cooling savings would win no?
Re: (Score:2)
Because I would guess, that the other things get much more expensive. I would guess, that few personell would like to live in some remote Alaskan or Canadian village, so you will have to pay them more, if you can even find some. Then you need a lot of power, I somehow doubt that that is available so easily either. Next is the problem with connectivity. A single connection is not exactly a good thing for a datacenter, you want to have redundancy. Also you have to move the equipment, whenever new hardware is
Re: (Score:1)
I've wondered why they don't put datacenters in old textile industry centers like Lowell, MA and Augusta, GA. Both of these places have canals that once supplied the mills with running water that drove turbines. You could rebuild the turbines to generate electricity and draw water off the canal for cooling. Plus mill towns tend not to be too far away from fiber, if there isn't already enough capacity there.
If someone has a couple million in venture capital to spare, I would like to attempt a project like
Re: (Score:1)
Re: (Score:2)
Yeah, like Manitoba Hydro.
http://hydro.mb.ca/ [hydro.mb.ca]
Winter gets you cold, cold, cold temperatures. Hydro power here costs you 5c/kWh and much cheaper for larger users. Want it even cheaper, get up north to Thompson - closer to the source.
Yet, no large data centers here. And in a "town" of 600,000+ people.
Re: (Score:1)
Plus Canada has lots of green hydro electricity. And data stored in Canada is exempt from the US Patriot act.
Easy solution (Score:1)
this is "news"? (Score:1)
Easiest savings come from free cooling... (Score:1)
Re: (Score:2)
Outside air can be very humid at night... not sure how high humidity it's safe to run a datacenter at though.
Re: (Score:2)
No, you are just ignorant like vast majority of people. It is called "RELATIVE HUMIDITY", NOT humidity.
100% saturated air at 0C is dry when heated to 20C = about 26% @ 20C.
http://einstein.atmos.colostate.edu/~mcnoldy/Humidity.html [colostate.edu]
100% humid air at -30C, raised temperature to 20C gives you humidity of 2% @ 20C. That is DANGEROUSLY TOO LOW for data centers. You want about 30% AFAIK otherwise you risks static charge problems with people zapping servers and servers even zapping themselves.
Re: (Score:2)
I know about relative humidity.
I have no idea where you're finding -30C air... maybe you should sell it.
I'm in southern California; on a cool summer's night, it might get down to the high 50's Fahrenheit with 70% or higher humidity.
60% (Score:2)
not sure how high humidity it's safe to run a datacenter at though
60% seems to be the common recommendation among datacenter humidifier vendors, even those who could sell more gear by changing that number. Static sucks.
Re: (Score:2)
Cool. Turns out it's been over 80% humid at night here lately. (so Cal, 63F and 84% humidity at the moment.)
Re: (Score:2)
Be careful though, you can get from nice humidity to condensing humidity in short order. Static sucks for servers, but drops of water can be worse!
Re: (Score:2)
I know, that's what I was trying to point out to jhw539 above; that outside air at night may be TOO humid to use for pumping into the datacenter, and I'm guessing that dehumidifying it may not be much cheaper than running the A/C instead.
And I was thanking you for the 60% ref.
Personally, I think more swimming pools should be used as liquid cooling to improve the efficency of air conditioners... you heat the pool for almost free, and your a/c runs cooler.
I don't know how bad the corrosion would be, but some
"Easy" (Score:2)
Get the self contained Toshiba 5MW reactor, build data center around it.
For desalination plant design, see above.