Efficient Supercomputing with Green Destiny 193
gManZboy writes: "Is it an oxymoron to have an efficient supercomputer? Wu-Chun Feng (Los Alamos National Laboratory) doesn't believe so - Green Destiny and its children are Transmeta-based supercomputers that Wu thinks are fast enough, at a fraction of the heat/energy/cost, according to ACM Queue." 240 processors running under 5.2kW (or less!) is nothing to sneeze at. The article offers up this question: might there be other metrics that might be important to supercomputing, rather than relying solely on processing speed?
Holy crap! (Score:5, Funny)
Indeed... (Score:1)
Re:Indeed... (Score:2, Informative)
Re:Indeed... (Score:1)
Re:Indeed... (Score:1)
Re:Indeed... (Score:1)
Re:Indeed... (Score:2, Interesting)
The heat output is precisely the same as the power input. All electrical power used by the PC is eventually converted into heat in the room, so a 450W PC consumes 450W of electricity and provides 450W of heat.
Incidentally, if you have a 500W heater in your room, you could replace it with a 500W PC for no extra electrical cost, and the same effect in terms of keep
Re:Indeed... (Score:2)
Re:Indeed... (Score:2)
Re:Indeed... (Score:2)
Oh, wait. I thought you said as much heat as the average PC enthusiast. Never mind.
Perhaps.... (Score:3, Insightful)
Well, IBM does have enough clout with S/W shops... (Score:2)
Paul B.
Re:Perhaps.... (Score:2)
-buf
ObDisclaimer: I'm an IT geek for RLX
Old age question for a new generation (Score:5, Interesting)
I think the question is a bit naive though as everyone knows a hundred software tools to rate performance of CPUs rather than just relying on MHZ.
Nick Powers
Re:Old age question for a new generation (Score:3)
Re:Old age question for a new generation (Score:4, Insightful)
Just look at cars - time was the only thing many people would look at is cubic inches or horsepower. Now most people who buy a car are more concerned with other features - passenger comfort, style, efficiency. I would guess this is a shift from car-oriented people buying cars to everyone buying cars as they became more of a necessity.
Computer manufacturers are only just starting to see this, making smaller, quieter, cooler-running machines. Hopefully they'll continue to look at what their customers actually need rather than simply putting out chips with higher clock speeds.
Re:Old age question for a new generation (Score:2, Interesting)
What ever happened to safety?
Computer manufacturers are only just starting to see this, making smaller, quieter, cooler-running machines. Hopefully they'll continue to look at what their customers actually need rather than simply putting out chips with higher clock speeds.
You are talking about computer manufacturers as if they are all in the same business. It's the chipset manufacturers that
Come again? (Score:2)
I don't know anything about you, but I'm now absolutely certain you're not a resident of the United States.
Re:Old age question for a new generation (Score:2)
Exactly.
Looking for cars, of course I can't help but look first at "The One Number", how many horsepower (BTW, OT, does the rest of the world measure automobile performance in Watts?)
These days, though, I'm very impressed with cars that have a high ratio of horsepower to gas mileage.
In the metric system, I guess that would reduce to kilograms per cubic second.
Maybe there's a similar ranking for peak torque vs gas mileage; it would be interesting to see a ranking of cars on this basis...
Re:Old age question for a new generation (Score:2)
Imagine... (Score:1)
Speaking of which, why hasn't anyone made an OpenMosix cluster-in-a-box yet?
Re:Imagine... (Score:2)
WTF? (Score:3, Insightful)
Re:WTF? (Score:1, Informative)
Re:WTF? (Score:5, Interesting)
Modest supercomputers are used in the military on airframes. Power consumption is important for at least two reasons. First is the wattage and power draw. Second, and more subtle, it that the cooling requirements while flying at high altitude become more important than simple fan noise. Pentiums burn up no matter what you do. PowerPCs@10Watts with conduction cooling will survive.
Re:WTF? (Score:2)
I would have thought that abundant cool air supply wouldn't be a problem for high-flying aircraft. Duct in outside air and use a simple heat exchanger if worried about airborne contaminants.
What's the outside air temp at 10,000 feet? Manage it with hot exhaust gas if a steady temperature is required. Where's the worry, freezing the chips?
Re:WTF? (Score:2)
Re:WTF? (Score:2)
Seriously, the power processors dissipate determines how close together they can be packed, and that average distance DOES set an ultimate limit on the latency of communications between the processors (speed of light, you know... ?
Would native VLIW be better? (Score:2)
If it would only be a few percent, it wouldn't be worth it. Transmeta has picked the game they want to play, and it would be a big deal at this point to engineer a special version of their chips that make it possible to run native VLIW code.
I'm guessing that typical scientific processing involves a lot of loops that run many iterations, which is the ideal situation for the
Does it really save that much power? (Score:3, Interesting)
Applying that here, the little VIA chips run at roughly the speed of a Celeron 500 or so, I'd say something like an AMD Athlon 3GHz would be just about as fast as about 6 of the VIA chips. So you are still saving some power, but as not as much as it would seem as first, as you need many low power chips to equal the speed of one faster chip. Not to mention power consumed by having more motherboards, network cards, switches, and other associated hardware.
Something to really look at is the cluster of G5's. The G5 chips use a lot less power than their x86 counterparts. I bet that cluster of G5's is probably right up there in terms of processing power per watt as this VIA super computer. And it's way more cool to boot.
Re:Does it really save that much power? (Score:1)
Re:Does it really save that much power? (Score:5, Informative)
1) A Nehemiah core C3 runs really close to the same performance of a comparably clocked Celeron, with the same general power consumption of a Samuel2 core (For those that don't know, part of how VIA's chip originally got it's low power is that the FPU was underclocked by a factor of 1/2). It's a nice chip overall, but it's not really intended (nor are they USING it that way) for scientific or gaming applications even though you can use it for that. The C3's winning usages is in something like a media PC, workgroup servers, and embedded systems where you need low power consumption, relatively low cost, and relatively high performance compared to other x86 embedded solutions.
2) The Crusoe and similar chips are very fast executing VLIW CPUs (very much like the Itanium...) that have code morphing that translates x86-32 instructions into comparable sets of instructions for the VLIW CPU- in fact it's very good at doing this sort of thing. The reason it's less desirable with a desktop or gaming application is that you're exceeding the VLIW code cache regularly, meaning you have to keep recompiling the x86 instructions into the native VLIW ones. For a scientific application, the same task gets executed time and time again and usually ends up with most, if not all, of the code in the pre-morphed code cache. At that point, you're now in the high-performance domain with very little power consumption. The Crusoe in this application would consume less power than the G5 and run just as fast. (Check the article that you're commenting on...)
Do some thinking outside of the box here, what's good or great on a desktop machine isn't always the optimal choice for supercomputing clusters or HA clusters. Depends on a bunch of factors, including what you're going to be running on the systems in question and what kind of environmental conditions you're going to be facing.
Re:Does it really save that much power? (Score:2)
Considering that FPU performance is particularly useful in scientific apps, this makes VIA chips nearly worthless there. Epia boards are cute & fit in toasters, but not really practical or cost effective.
Still trying to read past the first page (Score:2)
If you do the math with X (10,280 instead of 13,880 performance, 1000sq instead of 21,000sw, and 800kw instead of 3,000kw) you get a 337 fold increase in performance per square foot, rather than 65, and an 832 fold increase in performance per Watt, rather than 300 fold, vs the Cray.
And I don't know what the numbers for the Transmeta solution is.
G5 = HOT (Score:3, Interesting)
Re:Does it really save that much power? (Score:1)
Hmm... (Score:1)
Overall performance is much more important than efficiency. While efficiency is commendable at all computing levels, if efficiency is a very important aspect, then a supercomputer is probably not for you.
Do the math (Score:5, Interesting)
The other point is: how expensive it is to support a cluster ? Not only the energy consumption, but also the infraestructure. It is pretty darn difficult to keep a thousand processors cold. You may need a special building, special power supply for it, etc.
A final point: as far as I know, the rule of thumb is that the floating point performance with these energy efficient processors is of the same order of magnitude as regular processor, may be a factor 2 difference.
You do the math ... :-)
Re:Do the math (Score:3, Interesting)
I really can't tell now that the site is slashdotted. The CPU in this case can't be that much of a burden if they run around five watts. I am curious if 80% of the heat generated here is simply networking.
Re:Do the math (Score:1)
Re:Do the math (Score:2)
For one thing, a bottleneck in supercomputers is in most cases the network. In this regard, dropping some per/node performance might not affect the overall performance for applications that need intensive interprocess communication.
OTOH, with faster nodes you need fewer nodes and thus also less network traffic.
IIRC, the green destiny uses plain ethernet for networking, so it won't be able to compete with higher end cluster interconnects anyway (ethernet latency kills performance for many applications).
Re:Do the math (Score:2)
You're saying it. At the (undisclosed southwestern desert location) high-performance computing center on campus they have problems during the summer. On some of the hottest days, they have to stand outside with a garden hose & spray down the cooling-coils on the main AC unit, or else the whole system goes down. The system can handle the load, it just can't get the heat out of the system when it's 100F outside.
Bigger clusters (Score:2)
Laptop/Desktop/Handheld Supercomputers? (Score:2)
Especially with the new wave of Media Center based PCs...small small machines that are very powerful....is THIS the future of servers? Perhaps in a few years my web pages will all be served up from something like a handheld PC, with several processors and always-on WiFi? The possibilities are endless, but I see this DEFINITELY making it into laptops of some creed...those ultra-high-p
supercomputers vs man's only finite resource (Score:5, Insightful)
Time is the only truly finite resource from a human perspective. As technology has progressed, distances have been conquered, vast energies harnessed, but old Father Time is still inescapable.
As a result, we place great value on just how much time is taken to accomplish anything.
Re:supercomputers vs man's only finite resource (Score:1)
With advances in quantum physics, who knows how long this will be the case?
Money is finite too (Score:2, Insightful)
Re:Money is finite too (Score:1, Funny)
moving on...
Re:Money is finite too (Score:3, Interesting)
Re:supercomputers vs man's only finite resource (Score:2)
Time is the only truly finite resource from a human perspective. As technology has progressed, distances have been conquered, vast energies harnessed, but old Father Time is still inescapable.
It seems that at least one person [johntitor.com] claims to have conquered time.
Nano-ITX (Score:3, Interesting)
[ see image here: peertech.org/hardware/viarng/image/nano-itx-c5p.j
With 5,200 Watts for Green Destiny, you could use 433 boards these boards for the same power consumption.
The on chip AES is clocked at 12.5Gbps, Entropy at 10Mbps (whitened). Thus you would have
422Ghz of C5 processor power
5.412TB/s of AES (yes, terabytes)
4.22Gbps of true random number generation.
Yeah, these are really rough estimates, but that is a long of bang for your kilowatt buck no matter how you slice it.
With a cutting edge P4 approaching 100W the efficiency of these less powerful but fully capable systems will become increasingly attractive.
I would not be surprised to find bleeding edge processors relegated to gamers and workstations as most computing tasks start migrating towards small, silent, low power systems that simply *work* without eating up desk space, filling a room with fan noise and driving the electricity bill higher with continuous 100's of W draw.
Too bad it's not out yet... (Score:1)
Nano-ITX was annonced only a month and a half ago, and hasn't been released yet. So at least wait until the end of the year when they get it out before suggesting building beowulf clusters out of it.
Re:Too bad it's not out yet... (Score:2)
see also the SMP dual C5P mini-itx:
Small is beautiful @ extremetech [extremetech.com]
more important metrics that just speed? (Score:1)
Yes, people often consider flops/watt to operate, and flops/dollar to buy.
Speed alone means nothing. All these atoms in my apartment can do billions of operations per second, but they can't even play mp3s.
well hey (Score:1)
now we have palm pilots and watches that can store data (see the usb wrist watch)
so, really, a supercomputer that doesnt use that much energy isnt impossible.
anything's possible, one just has to break through the set barriers technology has made. if no one did that, we still would be sitting around
Times change (Score:3, Insightful)
from the so-obvious-it-hurts dept. (Score:3, Informative)
Um, yes?
Do the comparison with VT X instead of ASCII Q (Score:4, Informative)
Of course I dunno the numbers for the Transmeta solution yet!
NOP like there's no tomorrow! (Score:2)
. . . might there be other metrics that might be important to supercomputing, rather than relying solely on processing speed?
No. I have a rock that can sit and do nothing, consuming considerably less than even 5.2kW. You can talk efficiency and bang-for-buck all you like, but if you don't benchmark faster than (roughly) 100 common desktop machines, you don't get to call yourself a supercomputer.
Re:NOP like there's no tomorrow! (Score:2, Insightful)
That only shows how timely the definition of a supercomputer is. 100 common desktop machines are very uncommon and obsolete 3 years from now.
I think energy efficiency (MOPS/Watt) is a very relevant metric. The reason why my PDA cannot do wideband software radio or anything that needs lots of GOPS is energy-efficiency. If the same PDA could carry 100 XScale processors instead of 1 with the same battery lifetime, I'm sure we'll have applications for it in no time.
Re:NOP like there's no tomorrow! (Score:2)
That only shows how timely the definition of a supercomputer is. 100 common desktop machines are very uncommon and obsolete 3 years from now.
Right. I recall Apple making a big stink a few years back about being a desktop supercomputer when they hit 1 GFLOPS, or whatever the benchmark was that initially established the first supercomputers. What makes a computer "super" while Moore's law is still being met will definitely change over time.
I think energy efficiency (MOPS/Watt) is a very relevant m
Do companies care? (Score:2)
The article offers up this question: might there be other metrics that might be important to supercomputing, rather than relying solely on processing speed?
Yeah. Does the supercomputer do what the customer needed it to do? Nobody in the world lays down money for a "supercomputer" these days so that they can be the fastest kid on the block ... or at least they shouldn't. Ostensibly, there are massive amounts of computing work that they need done, and they need something that can do it in a reasonable amo
Imagine reading the beowulf mailing list! :-) (Score:1, Informative)
Memory Speed (Score:3, Interesting)
If its even just 512Mb of Gigabit ethernet and assuming 100% performace it would still take 5 seconds which is many orders of magniture. Just look at SGI machines which use NUMA and their Cray-Linux are 3.2 TeraBytes (bytes not bits). Now thats how you want to shift data
Rus
Re:Memory Speed (Score:2)
Efficiency (Score:1)
I thought heat was the real poison of "ultimate" computing...
So it seems likely computers will move towards those limits and become 'greener' with respect to how much energy they use...
To do otherwise would be counterproductive in terms of both efficiency and ecology.
That's needed given how much energy the US is using [cornell.edu]
Not a bad thing - but I wonder when green will move towards a technology that means less polluting in terms of hardware that gets
Being green is important (Score:4, Funny)
-Shane
Re:Being green is important (Score:2)
depends on the TASK... (Score:3, Insightful)
Less heating=Denser packing (Score:2, Interesting)
Re:Less heating=Denser packing (Score:2)
You're right about heat and density. Heat and density are directly associated. It's the reason 1.6U Evelocity nodes were chosen for Orange [linuxbios.org] instead of
Green Destiny? (Score:1)
Yes, yes, those numbers are impressive but can it be used to destroy other weapons and conquer the Chinese underworld in the hands of a rebellious Manchurian girl? (Reference: Crouching Tiger Hidden Dragon)
Why does a supercomputer need x86 compatibility? (Score:3, Interesting)
Re:Why does a supercomputer need x86 compatibility (Score:2)
Probably, if you have an optimizing compiler for exactly that chip. On the other hand, the code-morphing engine is already quite efficient, and by keeping the internal ISA "hidden" or "secret", each new model can have a completely new ISA, and they never have to worry about backwards compatibility.
Re: (Score:1)
Re:Steam powered PC (Score:2)
The short answer? (Score:1)
Faster, ever faster!
Full Formatted Text (Score:5, Informative)
Making a case for Efficient Supercomputing
From Power [acmqueue.com]
Vol. 1, No. 7 - October 2003
by Wu-Chun Feng, Los Alamos National Laboratory It's time for the computing community to use alternative metrics for evaluating performance.Motivation
A supercomputer evokes images of big iron and speed; it is the Formula 1 racecar of computing. As we venture forth into the new millennium, however, I argue that efficiency, reliability, and availability will become the dominant issues by the end of this decade, not only for supercomputing, but also for computing in general.
Over the past few decades, the supercomputing industry has focused on and continues to focus on performance in terms of speed and horsepower, as evidenced by the annual Gordon Bell Awards for performance at Supercomputing (SC). Such a view is akin to deciding to purchase an automobile based primarily on its top speed and horsepower. Although this narrow view is useful in the context of achieving performance at any cost, it is not necessarily the view that one should use to purchase a vehicle. The frugal consumer might consider fuel efficiency, reliability, and acquisition cost. Translation: Buy a Honda Civic, not a Formula 1 racecar. The outdoor adventurer would likely consider off-road prowess (or off-road efficiency). Translation: Buy a Ford Explorer sport-utility vehicle, not a Formula 1 racecar. Correspondingly, I believe that the supercomputing (or more generally, computing) community ought to have alternative metrics to evaluate supercomputersspecifically metrics that relate to efficiency, reliability, and availability, such as the total cost of ownership (TCO), performance/power ratio, performance/space ratio, failure rate, and uptime.
Motivation
In 1991, a Cray C90 vector supercomputer occupied about 600 square feet (sf) and required 500 kilowatts (kW) of power. The ASCI Q supercomputer at Los Alamos National Laboratory will ultimately occupy more than 21,000 sf and require 3,000 kW. Although the performance between these two systems has increased by nearly a factor of 2,000, the performance per watt has increased only 300-fold, and the performance per square foot has increased by a paltry factor of 65. This latter number implies that supercomputers are making less efficient use of the space that they occupy, which often results in the design and construction of new machine rooms, as shown in figure 1, and in some cases, requires the construction of entirely new buildings. The primary reason for this less efficient use of space is the exponentially increasing power requirements of compute nodes, a phenomenon I refer to as Moore's law for power consumption (see figure 2)that is, the power consumption of compute nodes doubles every 18 months. This is a corollary to Moore's law, which states that the number of transistors per square inch on a processor doubles every 18 months [1]. When nodes consume and dissipate more power, they must be spaced out and aggressively cooled.
Figure 1
Without the exotic housing facilities in figure 1, traditional (inefficient) supercomputers would be so unreliable (due to overheating) that they would never be available for use by the application scientist. In fact, unpublished empirical data from two leading vendors corroborates that the failure rate of a compute node doubles with every 10-degree C (18-degree F) increase in temperature, as per Arrenhius' equation when applied to microelectronics; and temperature is proportional to power consumption.
We can then extend this argument to the more general computing community. For example, for e-businesses such as Amazon.com that use multiple compute systems to process online orders, the cost of downtime resulting from the unreliability and unavailability of computer systems can be astronomical, as shown in table 1millions of dollars per hour for brokerages an
Other Metrics... (Score:1)
yes, cost...
New trend in computing. Vector processing (Score:3, Informative)
Re:New trend in computing. Vector processing (Score:2)
Re:New trend in computing. Vector processing (Score:2)
This makes little sense to me (Score:2)
It would seem to make more sense to shut down and start up individual nodes for power saving. A supercomputer has relatively little down time and most of the time jobs come in large batches, you can shut down or heat up additional nodes as needed. It should be relatively easy to implement this kind of power saving inside single "computers" with many processors, and of course absolutely trivial to do it with a cluster. Especially if you use WoL NICs.
Re:This makes little sense to me (Score:2)
In some applications, you don't have any disks, though clearly that isn't for everyone. Mostly, if you buy good drives, then you can be sure they will keep starting up. Or, you can buy cheap redundant drives, and RAID everywhere. Or, what I would do, is put all the drives in sleds so when one dies you can just slide another one in.
You can naturally wake machines up periodically to test their hardware.
SMART will give you a good idea if many drives are going to fail, before you have shut down and faile
"Fast enough" != "Supercomputer" (Score:2)
Unless their architecture actually hits the Top Ten [top500.org], I'm not going to be impressed that it's overcoming its handicap. Unless you're running a Special Olympics [specialolympics.org] for computers and "everyone's a winner."
StrongARM (Score:2)
--
Real computer scientists despise the idea of actual hardware.
Hardware has limitations, software doesn't.
It's a real shame that Turing machines are so poor at I/O.
Re:StrongARM (Score:2)
The only reason FPUs were not integrated in StrongARM is that there cost (in terms of Watts) is not justified by the small benefit to performance (in terms of MIPS/FLOPS). Aparently, FPUs, in their current form only make se
Power Efficiency (Score:2)
For supercomputing, I would imagine that something like SPECfp/watt or SPECrate/watt would be a decent metric.
If your limitation is a finite power budget, then you pick the most highest perf/watt CPUs.
P4 3.2 EE = 18.44 SPECfp/Watt (80 watts)
Crusoe = ?? No performance numbers published, but I'll bet you it's lower
Building larger caches (which can be made low power) is a good way to acheive high power/performance efficiency.
Duh (Score:2)
Not to sound flippant, but...duh. Okay, I know I sound flippant. But seriously, why has it taken so long to realize that processing speed is not of the utmost importance? It's like saying one car is better than another because has a top speed of 180MPH and the other 174MPH, ignore that the "slower" car gets 30% better mileage. There's such a thing
That is all fine and good, until.... (Score:2)
They modified the code morphing software? (Score:2)
That's not exactly trivial. Does anyone know wnything more about this? What are the trade-offs?
And, will thi
Re:Green Destiny (Score:2, Interesting)
Re:Green Destiny (Score:1, Informative)
Re:Green Destiny (Score:2)
5.2kW cannot be sucked out of "a normal building power strip." And you are sure as heck going to notice 5.2kW of heat, and the regular everyday HVAC is most definitely not all it requires. "Uncooled ordinary room" my ass.
Re:Green Destiny (Score:2)
Re:Green Destiny (Score:2)
Re:I can think of one (Score:2)
Let me get this straight, you're comparing overall computing performance of a chip based solely on its clock speed? You must be new around here.
Re:I can think of one (Score:1)
Have you ever tried picking one of those things up? I have. I worked on the G5 cluster. Those SOBs were heavy. Nice to look at, but they suck to bring to LAN parties.
Speaking of G5's and recent price drop (Score:1)
I think apple might have to buy back some stock or give discounts to retailers. It's pretty vicious to drop down a system that completely obsolete's the retailers stock so soon. Vicious to the retailers at least
I want a G5, but i dont have a place to put it. My wife has officially limited the number of computers to 5.
Re:green destiny? (Score:2)
I could just see the DEA raiding Los Alamos after getting reports of computer scientists blowing entire afternoons on "their smoking Green Destiny"