Truly Off-The -Shelf PCs Make A Top-500 Cluster 231
SLiDERPiMP writes: "Yahoo! News is reporting that HP created an 'off-the-shelf' supercomputer, using 256 e-pc's (blech!). What they ended up with is the 'I-Cluster,' a Mandrake Linux-powered [Mandrake, baby ;) ] cluster of 225 PCs that has benchmarked its way into the list of the top 500 most powerful computers in the world. Go over there to check out the full article. It's a good read. Should I worry that practically anyone can now build a supercomputer? Speaking of which, anyone wanna loan me $210,000?" Clusters may be old hat nowadays, but the interesting thing about this one is the degreee of customization that HP and France's National Institute for Research in Computer Science did to each machine to make this cluster -- namely, none.
Can you imagine? (obligatory) (Score:5, Funny)
-Berj
Re:Can you imagine? (obligatory) (Score:1)
harmony (Score:1)
Pentium 1 cluster (Score:1)
Re:Pentium 1 cluster (Score:1)
Re:Pentium 1 cluster (Score:1)
Now, just start sending those donations...
Re:Pentium 1 cluster (Score:1)
What was even cooler... (Score:4, Funny)
$210,000 ?? (Score:5, Interesting)
Power consumption of old CPUs (Score:2, Insightful)
Is there anything like a MIPS/Wh rating for CPUs? (Would thermodynamics dictate a certain minimum?)
With a seperate power supply and hard disk per CPU (i.e. complete box) I would imagine that old PCs generate a *lot* of heat per CPU cycle.
Has anybody done measurements/calculations on this?
Re:Power consumption of old CPUs (Score:3, Informative)
Also, I don't think you'd get much useful stuff done with early Pentiums and 486. Consider that a P4 2 GHz has 20 times the clock speed and probably does twice as much per cycle, so it's ~40X faster. Now, if you connect 40 P100 together, unless your problem is completly parallel (like breaking keys, as opposed to most linear algebra), you're going to lose at least a factor of 2 there. This means that in order to equal 1 P4 @ 2 GHz, you'll need almost 100 Pentium 100 Mhz. This means that 10 P4 would be like a thousand Pentiums. At these numbers, it's going to cost so much in networking and power...
I'd say (pure opinion here) the slower you'd want to have today is something like a Duron 1 GHz and the best MIPS/$ is probably with a bunch of dual Athlon 1.4 GHz (A dual is not cheaper that 2 single, but you get more done because of parallelism issues).
Re:Power consumption of old CPUs (Score:3, Informative)
If you really want a cpu that does a lot in a single cycle, look at the IBM POWER series. IIRC, on the floating point side, a 2xx MHz POWER III is darn not too far from an Alpha 21264 at 733 MHz. And now there are 1.1GHz and 1.@GHz POWER IV chips, in the new IBM p690 machines. I don't know how they compare to the POWER III per cycle, though, because the POWER IV opens a whole new (good) can of worms.
-Paul Komarek
Re:Power consumption of old CPUs (Score:2)
For "unoptimized" applications, that may be approximatly true. However, if you're going to build a cluster, you're also going to optimize your code for it. What kills the P4 is branch misprediction. However, by carefully writing your code, you can avoid most of these problems. Also, most of the big clusters are for numerical code, for which branch predictions does well (plus you can do lots of loop unrolling).
Another thing that the P4 (and PIII) has is SSE. On a P4 2 GHz, you can do theoratically 8 gflops. In practice, if you write good code, you'll bet between 1 and 2 gflops. On a plain pentium, the FPU is not pipelined, so a P100 (I'm guessing here) probably has a theoretical maximum of ~25 mflops with a performance for real code around ~10 mflops. That means the P4 is probably 100-200 times faster at floating that a P100.
Of course, you're right in saying that other architectures are probably faster that P4.
Re:Power consumption of old CPUs (Score:2)
I've never tried optimizing for SSE, but someone in lab did once. He reported higher performance when doing his computations element-at-a-time than vector-at-a-time. His conclusion, for his particular application, was that memory latency was killing SSE. He was better off doing lots of work on a few numbers, than he was doing fancy stuff to a lot of numbers with SSE. On the other hand, some people have had some luck with SSE optimized FFTs, or so I've heard.
At 2GHz, I'll bet that you're better off doing element computations than vector computations because of the radical difference in memory-versus-processor performance, if the P4's L1 take more than a cycle to feed the registers. Otherwise, do whatever fits in L1 and can be prefeteched -- like elementwise computations on long vectors. Anyone have any real or anecdotal evidence to refute or support this theory?
In the end I think that platform-specific optimizations are a waste of time for research code. I seem to remember some people eventually including hooks in BLAS or LAPACK to allow the user to specify cache sizes; and FFTW does some runtime experiments for optimization. But my guess is that, overall, SSE, 3DNOW, and even AlitVec are irrelevant to most computer researchers. I'll bet their highly relevant to most embedded engineers or many robotics researchs, i.e. people targetting specific hardware.
-Paul Komarek
Re:Power consumption of old CPUs (Score:2)
Maybe his problem was really special, but most likely he didn't know how to write SSE code. First, if you write your code correctly, the worse you can do is a bit better than the x87, because you can use SSE with scalar and take advantage of the linear register model (as opposed to stack).
The only time I've converted some code for SSE, I got a 2-3x improvement in speed. There is one thing you really have to be careful when writing SSE code: ALIGNE EVERYTHING TO 16 BYTES. That's very important. The "movaps" (move aligned packed single) is more than twice faster than the "movups" (unaligned) when data is unaligned (movups on aligned is not too bad). That makes all the difference. Also, sometimes you need to change the way you do your computations.
For the case I have (optimizing a speech recognition engine), just by changing the order of of the loops (inner vs. outter), we got a 2-3x improvement (still with x87) because of increased L1 hit rate. Then when switching to SSE, it ended up to a 5x improvement over the original code. Had I just re-written the code in SSE (with cache optimization), the gain would have been around 25%, because memory would still be the bottleneck.
As for libraries not being optimized, just look at FFTW (www.fftw.org) or ATLAS (www.netlib.org/atlas) and you might change your mind.
Re:CPU not the limiting factor (Score:2)
No way. Weather sinulations involves lots of linear algebra on huge matrices. When you parallelize that, you need a lot of communication between the nodes. With a cluster of P100, communication will kill you right there (10x of so penalty). It's not that much the network bandwidth, but the latency. Weather simulation is one of the hardest problems to parallelize and that's why until recently, SMP was prefered to MPP (and of course clusters of small workstations).
As for the memory bandwidth, depending on the problem, sometimes the L1 is really effective.
Practically anyone? I think not. (Score:3, Insightful)
Unless "practically anyone" has the funds, the storage room, and the manpower to maintain this monstrosity, there is nothing to worry about.
And even if anyone could build a supercomputer, what's there to worry about? We don't live in the "War Games" world where supercomputers play chess, tic-tac-toe, and start nuclear wars for fun.
Re:Practically anyone? I think not. (Score:1)
Re:Practically anyone? I think not. (Score:3, Informative)
Re:Practically anyone? I think not. (Score:3, Insightful)
Knowing nothing about it, I would nonetheless guess that it's rather non-trivial.
Keep in mind that nukes were invented without the aid of a Beowulf or a Cray.
Nukes must also be TESTED (Score:1)
Re:Practically anyone? I think not. (Score:2, Funny)
Re:Practically anyone? I think not. (Score:1)
Or terrorists could use the Athlon cluster for a more insidious plan. Simply plug them in, turn them on and with the juice they draw, black out any major metropolitan area of their choosing, oh the horror! *g*
Re:Practically anyone? I think not. (Score:1)
Unless "practically anyone" has the funds, the storage room, and the manpower to maintain this monstrosity, there is nothing to worry about.
It's only 256 machines, how hard could it be to steel(steal?) 256 computers. Rip off a couple of Gateway (are they still in business) and Dell trucks, maybe rob a few schools. Now that you've gotten (by fair means or foul) a super-computer that is rated in the top 500, what would you like to do with it? How about any brutt force crack that you feel like (it's not like you need some neat algorthyms with that much firepower)? How about generating insane encryption levels that FbI and CiA can't crack?
Of course, about 256 people with laptops and wireless modems, all running mandrake, sitting in the train station solving tough mathematical problems just annoy people....
Putting them to work. (Score:5, Informative)
Data transfer and message passing are such a big issue I belive the most important developments are in the networking topologies and hardware for these environments.
That said, I still want one in my basement
Re:Putting them to work. (Score:1)
Yeah. Me too.
This even more strongly underscores the current state of the art: we have astonishing hardware. OK, now what do we do with it? Other than grumble about Windows XP memory requirements? :-)
I wouldn't mind finding out. I bet it will be cool.
...laura
Re:Putting them to work. (Score:1)
That'd be good in the winter, but, hey, it's going to be really hot at your place next summer!
This just goes to show you (Score:4, Insightful)
Now all we need are ways of getting local connections significantly faster (Did someone say Gig Ethernet) to allow faster communication between the nodes and we will be able to scale beyond several hundred and break the top 100. I hear 1gig NICs will be falling in price to under $100 US retail soon...
How fast do you connect to your cluster ?
Re:This just goes to show you (Score:1)
Re:This just goes to show you (Score:1)
Re:This just goes to show you (Score:3, Insightful)
I hear Gigabit switches won't be...
Re:This just goes to show you (Score:4, Informative)
The real issue is how parallel-efficient your algorithms are. We do molecular dynamics (MD) on large clusters, and we can get away with slow networks because each node of the cluster has data that is relatively independent of all other nodes - only neighboring nodes must communicate. If you have a case (and most cases are like this) where every node must communicate to every other node, it becomes a more difficult problem to manage. To deal with this, you need a high-speed, low-latency switch like the interconnects in a Cray. The only real choice for that is a crossbar switch, like Myrinet.
And Myrinet is tres expensive.
Re:This just goes to show you (Score:2)
Or is th elatter to problem-dependent to make this practical?
worry? (Score:3, Funny)
Should I worry that practically anyone can now build a supercomputer?
Yes, you should probably worry that practically anyone can build a supercomputer. But you could mitigate all that fear with the fact that not practically anyone can whip up software that takes full advantage of it.
Thank god there isn't any off the shelf "missile trajectory" software in the CDW catalog. you would hope that any society that can whip together motivated coders to write such code already has access to some pretty spiffy kit.
(yeah i said "kit"... and I'm from Chicago... I feel like such a wanker.)
On the preponderance of "Kit" (Score:2)
Yes, you heard me right. Cultural assimilation. The brits are sick of seeing Mickey Mouse and Donald Duck and the sexy chick from Enterprise on TVs all over the world, and they're going to do something about it.
The British invented the English language, and in many circles certain British accents are percieved as more sophisticated or upper-class. They're capitalizing on that by inventing slang terms - "kit" being among the forerunners - that other English-speaking peoples appropriate. Thus is begins.
Soon, British TV will move off of PBS, where it belongs. British computer games and hardware will surpass American in popularity. And there is nothing - absolutely nothing - we can do about it.
(In case you hadn't realized it, yes this is a joke. And yes, I know it's offtopic and will be moderated as such. But this was fun to write.)
Kit == Rig (Score:2, Informative)
No it isn't. That's just the only context in which you've heard it used (translation: you read too much Slashdot, and should get out more often). "Kit" is the British equivalent of the American "rig" when used in this context. It is not used specifically to refer to computers.
Thanks for the info (Score:2)
Of course, you know that makes the threat presented by the word even more insidious. If non-techies can use it well - I shudder to think of the potential for linguistic infiltration!
Re:On the preponderance of "Kit" (Score:2)
Your plan is futile.
Like most Americans, when I hear the word "kit" used in a technological context I instantly think of the popular 80's television show "Knight Rider" starring David Hasselhoff [davidhasselhoff.com].
Now if you were German...
Yes, it was = you're right (Score:2)
Re:Yes, it was = you're right (Score:2)
Yes there is (Score:2)
Hail Brittania!
An interesting project (Score:2, Interesting)
The ultimate Linux selling tool, every linux box in your company is a node in a cluster, add a few servers for extra speed, add a few computers to provide file I/O and backup capability, and you have one of the fastest supercomputers available to your company without having to spend an extra dime (everyone needs a desktop anyway). Can you imagine the extra cycles available for simulation, whatever when people start going home at 5 PM.
Re:An interesting project (Score:1)
Re:An interesting project (Score:2, Insightful)
That is a serious problem though, and one I assume Beowulf clusters will take care of, what if a node goes down in the middle of processing, how does the cluster respond to it ?
Re:An interesting project (Score:2)
We do use desktops at night to work with our render farm. Platform has some cool tools to work with for such environments. I have never tried LSF in conjunction with PVM or MPI but they have support for it, so I imagine it does pretty well.
Saddam Hussein's is faster (Score:1)
You'd think a top computer manufacturer would be able to beat out a part-time dictator from the third-world in gigaflops, but I'm thinking it was more a demonstration by HP that they're getting set to embrace Linux and shelve HP-UX.
Yahoo! news (Score:3, Funny)
Spare cycle approach is interesting... (Score:2)
Windows support makes a difference. Take a large company (10,000+ in a single location) that has some intensive projects. In this case, they could just drop the $210,000 (call it $750,000 with installation, support, etc.) and put it in a room.
However, a smaller shop, say, 50-250 employees, being able to install this software on the staff's machines. They rarely use their computers to capcacity, and can probably contribute 90% of the CPU 90% of the time. This approach could let people doing giant calculations do so cheaply.
The real question, however, is who needs that kind of horse-power. For those that need the horse power, is the savings with off-the-shelf components meaningful.
Its a tremendous accomplishment, and I wonder how much of the changes were new (vs. Beowulf clusters that we always hear about). However, if this fills a need, congratulations, its an impressive accomplishment regardless.
Alex
Re:Spare cycle approach is interesting... (Score:1)
There's a very good Greg Egan novel called Permutation City [amazon.com] that takes the concept of shared computing power a long way.
I don't have time to explain -- just wanted to plug a good book of (at least marginal) relevance to this topic!
Is this a story from 1993? (Score:1, Interesting)
[OT] Yahoo news posts (Score:1)
As soon as I see a blur (pop-behind) ad I quickly close the page I'm on in the naive hope that they track when people stop visiting their sites and don't leave via an ad...
okay um.. (Score:2)
also, who will pay your power bills?
i don't get this "drool factor" thing some people have for supercomputers... sure, they're cool and all, but they can do exactly nothing you would want or need to do on a day-to-day basis...
Re:okay um.. (Score:2)
Re:okay um.. (Score:2)
THAT'S the drool factor, so to speak.
Re:okay um.. (Score:2)
Okay, so I'm curious: (Score:3, Funny)
what scares me is... (Score:1)
this kind of hardware is out of date? unless i'm mistaken HP markets these e-PCs toward home users looking for light processing power, such as the ability to view web pages, read emails, and play solitaire. this looks more like a power-user rig, or something a gamer would have as a decent Q3A machine. how in the world could this hardware be obsolete? i guess i should replace the pentium III 933 i'm running because lord knows it just won't hold up to today's high powered apps! man it's almost a year old, i should start worrying...
pxg
Re:what scares me is... (Score:2)
It's at #385 in the list (Score:4, Informative)
The cluster is at #385
distributed supercomputer cluster (Score:2, Interesting)
Technically I can't see why this wouldn't be feasible. It would be beyond SETI and protein folding in that the 'control center' could change what problem was being worked on at any time. It may not be incredibly practical compared to setting up specific machines in a single large room, but it would be free and have a potential user base in the hundreds of thousands or millions.
Imagine: instead of the same SETI screen output time and again, you'd get a message on your SS saying "would you like to see what your computer is working on right now? How about high-pressure fluid dynamics in environment x?"
Max
Re:distributed supercomputer cluster (Score:1)
That being said, my approach has been to design a C/C++ like language that intrinsically understands multithreading and remote processing, but has a minimal standard library, and is by default intrinsically secure by design. People running the client software can allow default security privileges for anyone to run under, or can set up specific execution profiles for certain people or programs, controlling cpu load and even restricting the hours they share out their machine, etc. Basically, think SETI@Home, except with any program that might be written in the language automatically taking advantage of all available clients.
Trust me, there's a _lot_ of technological wrinkles to iron out in the language to make it safe to run on another person's machine without authorization, as this is the heart and soul of the project. Rather than reinvent the distributed processing technology for many different purposes and try to get many people to run several clients, it's better to get a single client that works for all sorts of problems. I think the time is now to prepare to capitalize on spare processing power. As we plunge further into the information age of cheap hardware and don't seem to require much more power to do the same tasks (unless you run Windows XP), the ratio of spare cycles to used ones will only increase over time.
The topology of a network of distributed clients over (potentially) slow connections makes it undesirable for solving some kinds of problems, though. What it's _not_ good for are things like converting video files from one format to another, like the HP I-Cluster is. Making it good at such tasks means slackening the security aspect a bit... And even then, slow transfers can invalidate even a fast remote CPU.
Of course, I'm always looking for funding. I'd love to make this a full time job in the future. Know anybody with loose purse strings? Didn't think so.
A cluster of Rippers (Score:1)
Any one else notice that this seems to be a very elaborate (and expensive) project just to bootleg encrypted DVDs?
Power Usage (Score:2, Interesting)
I was pondering the computrons per watt of a cluster such as this versus a real honest-to-Bob supercomputer (Something from Cray/Terra/SGI, for example). we can assume that each machine in HPs cluster uses probably 60-80 watts (because they're sans monitor), so youre looking at about between 1.2 and 1.8 kilowatt hours to power this thing. I'm not sure what a Cray TSE [cray.com] uses, but I have to think it's nowhere near that because of all the redundancy that PC clusters use (one Power supply, chipset, etc per Core).
Though, I'm sure if you can afford either a Cray or 256 PCs, you can afford the power bills, too. If you have to ask how much it will cost you, you can't afford it. But while CIP (Cluster of Inexpensive PCs) is cheaper, is it as efficient?
.
Re:Power Usage (Score:1)
Here's some info: http://www.cray.com/products/systems/t3e/> [slashdot.org]
Regarding efficient, in a word: no.
These systems, much like an OS/390 are written and built around a whole different theory of architecture. They are also often custom made for specific solutions. The system for 'missile guidance simulations' would have different cards, different chips, and way different software from the system for 'global thermo-nuclear warfare simulations'.
Cool to play on, though.
-WS
Re:Power Usage (Score:1)
with an efficient cooling system of network booting pcs (no hard drives, no moving parts except for fans) then you have a bank of systems which is not only incredibly cheap and off the shelf but efficient as well
what about... (Score:2, Interesting)
Re:what about... (Score:1)
Unless you know where you can get rackmounts case and powersupply for less than $50 each... you can't technically make a Beowulf cluster with them.
An idea for a business plan (Score:2)
Hey remember all those completely and hopelessly out of work Russian PhD CS grads sitting around and starving and writing strong crypto software for the Russian Mafia? You might even have heard that the Russian Mafia is always looking to explore new business ideas and strategy.
Well hell wouldn't this be a great business opportunity for both of them?Call it RMBM (Russian Mafia Business Machines), and then build cheap super-clusters and turnkey code for "specialized" clients. The possibilities are endless.
This is where you get them now: Support. You sell them the machines at a 25% markup and then charge a ridiculous annual service agreement.
From the presentation:
"Using "borrowed" Post-CCCP Mi-8TV assault/commando choppers RMBM support staff can be deployed to your corner of the desert in a matter of hours! Lets see IBM match that! Not even Larry Ellison and his personal Mig can touch that! (canned laugh track)"
I don't know, maybe not.
Downturn clustering (Score:2, Funny)
Re:Downturn clustering (Score:1)
Not to mention all of the dot-com fire sells.
Notice (Score:2, Funny)
Anyone who posts a comment containing the word "Beowulf" will be shot.
Including me.
Uh-oh.
Making a cluster work (OSCAR?) (Score:1)
Anyway... I just got a job working with the NCSA as part of a project called OSCAR [sf.net]. It's basically an open-source solution to the problem of creating a cluster. I'm part of the team working on documentation and training materials for people trying to implement OSCAR. I can say, from my own experience, that installing a cluster (even only with 4 PCs) is not a simple task. OSCAR is still young as far as software development goes, but it will do the job well once it is finished.
For those interested, OSCAR 2.0 is on its way soon.
Yahoo! got this story from CNET (Score:1)
amd for less (Score:2, Interesting)
for less money with amd procs/mobos/etc
1gig tbird $100
decent cheap amd mobo w/integrated vid/snd/net $100
256meg ram $25
15gig ide $50
floppy drive (needed??) $15
cdrom (needed??) $25
decent nic $20
cheap case $40
total $375
subtract 10% (due to quantity purchase) gives less than $350 total each
pay a bunch of college kids $10/hour
they'd build 2 machines/hour
so 125hrs total to build comps is $1250
$350 x 250 machines is $87,500
add in (8) good quality 32 port switches @ $200 each and you're up another $2k
add in 250ish cat5 cables for another $1k (who wants to make them, buy for $3-4 ea)
your total cost is way under $100k
or even better
use the new SMP durons, 1gig each
not much more $$ since durons are cheap
add like $50 for the 2nd proc (total $150 for 2 duron 1gig smps, unsure if thats reasonable pricing) and another $50 to mobo cost for dual smp mobo
thats $450 ea box
250 x $450 gives us $112.5k for the boxes
add in networking stuff etc
less than $125k prob
man i want to do this
need someone with $$ =P
--wayne =)
Re:amd for less (Score:2, Insightful)
Also why have a mb with integrated sound (and video) - this is a beowulf cluster, not a John Cage album...
graspee
again... Supercomputer? (Score:1)
So maybe by using e-pc's the peak performance went down some, but anytime you tie anything together in a clustered environment the sustained performance dies (not just takes a hit) too. Only way to make it hit close to peak is assign each node a process that doesn't require any interaction between processes/nodes. In that case, you wouldn't need to tie them together to make a top-500 cluster. Just assign some IP's, cross mount their filesystems and call it a morning.
Besides, government agencies (and scientific companies) are beginning to realize that when you cluster 500 boxes together, you're still administering 500 boxes. When real supercomputer companies make real supercomputers, you've only got to administer one computer. Maybe that's why the term "Supercomputer" isn't plural.
If clusters keep on being called supercomputers, you might as well call "the internet" a supercomputer too since it's an environment where lots of computers are connected and running processes that don't depend on another. "Wow! Look at the sustained power of all 5000 Counter-strike servers out there! It's a super-counter-strike-computer!"
Beowulf Politics? (Score:1)
Re:Beowulf Politics? (Score:1)
Mental note: Check for typos next time. lol.
DaveYawn. Another One. (Score:3, Insightful)
You know this Beowulf business is getting to be pretty staid and routine by now.
In fact, I'd almost say it would be newsworthy if there were any organization (university, company, govt lab) that had not yet built "a supercomputer from the COTS components".
What I'd like to see now is more metrics (some of which the article does, admittedly, reveal).
Don't get wrong. I love these clusters and want to use them. It's just that, in 2001, their mere existence is no longer as exciting as it was in the mid 1990s.
Now days, I care more about ease of use and ease of maintenance, taking the low cost of a Beowulf cluster as a given.
With the size of these clusters going up and the ratio of hardware cost to human time constantly decreasing, I'd be more impressed to see how a system with many hundreds of nodes was brought up in a short time, never rebooted for a year, even as 13 of the nodes developed variously problems and become unproductive members of the cluster.
"Clouds" of compute power (Score:1)
Anyone else reminded of A Deepness in the Sky by Vernor Vinge? In that story, IIR, one of the protagonists controlled a ubiquitous cloud of invisible compute "particles". Each particle was networked to the rest through its neighbors floating in the air around it.
Ok, so I thought it was a cool idea.
News worthy? (Score:2)
It looks like the wheel reinvented several times.
For cluster installs on several machines, use system imager [systemimager.org] .
For using and controlling a cluster of machines for various taskes, use LSF [platform.com] .
The number of machines is pathetic too
How sad. I need to bench mark our render farm (200+ boxes, 120 are dual 1ghz) and see what we can come up with. I know it is higher than that... and we have a smaller install for the industry.
I looked for info to spec our machines but I couldn't find any info.... any help?
Incorrect citation (Score:2)
No, it isn't. Yahoo! News is repeating a story which, if you'd bother to read the byline they wrote, was
The article [cnet.com] on CNET's site should be getting the Slashdot treatment, don't you think?
This Cluster and Beowulf not exportable? (Score:1)
Since you could now build a Supercomputer with OTC stuff (well, Im sure you could before today as well), will that make software like this (read: Beowulf) fall under export laws preventing exportation to countries like Iraq and China?
Any insites?
For those wondering about nuclear testing... (Score:5, Insightful)
While great at highly parallel tasks that require very little synchronization between threads (think code cracking), nuclear testing (and almost all other fluid dynamic problems) generally requires all of the cpu's to have high speed access to all of the memory. So one needs a huge shared memory system (think Cray or Sun StarCat).
And for this reason, I find the top 500 list to be a bit misleading in these days of massively parallel systems. Its great as a test of how many flops the system can crank out, but it does not take into account the memory bandwidth between the cpu's, and that is often more important than raw cpu horsepower.
Re:For those wondering about nuclear testing... (Score:3, Informative)
I find too that people assume that an "X" type of cluster will solve all problems, regardless of what they are. Each cluster type serves a purpose. Cray and then SGI spent time developing the Cray Link for a reason. Sun, IBM, HP and others have gotten into the game as well. Sometimes you need a ton of procs with access to the same memory, sometimes the task divides well.
I see this from almost the opposite side of the spectrum with rendering. To render a shot, you can divide the task amongst ignorant machines. They just need to know what they are working on. The cleverness goes into the managment of these machines. A place where the massively parallel machines would be nice is rendering a single frame. After the renderers initial load of the scene file and preping for render, the task can be divided amongst many processors on the same machine. To divide it beowulf style would throttle the network with the memory sharing over the ethernet ports.
So from my experience:
big data, long time
little data, long time
little data, short time
Re:Imagine... (Score:3)
I dunno. It's kinda lacking when you compare it to all the other Beowulf clusters we've considered.
Re:Imagine... (Score:2, Insightful)
I dunno. It's kinda lacking when you compare it to all the other Beowulf clusters we've considered.
Now, imagine a cluster of Athlon 1.4GHz machines doing the same thing ... now there's a drool factor, and probably cheaper to boot!!
Beowulf Athlons? (Score:1)
Re:Imagine... (Score:1)
Hofstadter's Law? (Score:1, Funny)
-AC 4eva! w00p! 4ll j00r b4s3!
Re:Technicly.. (Score:2, Informative)
An Apple is not a supercomputer.
RISC does not mean faster. It allows for simpler design which can lead to increased speed, but as we have seen, Apple have consistently failed to compete with Intel and AMD (not that they even make thier own chips...). CISC is actually a good idea, since with the huge speed differential between CPU and memory, and the introduction of cache, the bottleneck in any system is the memory bandwidth. Think for a moment : why did Intel add instructions to the x86 architecture in every iteration? Because its faster having one instruction doing something complex, than many simple ones, simply because of the reduced frequency of memory access. In todays computers, RISC doesn't mean anything, since memory, storage and network bandwidth is the bottleneck.
The moral of this story:
1: Don't believe Apple's advertising.
2: Don't believe what a Mac Zealot will tell you about RISC or some other claptrap.
3: Get ppc Mandrake if you're unfortunate enough to have actually bought a G4.
Yes, I use Macs. Daily. And I hate Apple. But my PHB is a Mac zealot. It frightens the hell out of me seeing all our company's work being stored on a Mac (OS9 (no pre-emption, memory protection, RAID, journalling, or anything you would want for a server...)).
Re: RISC (Score:2, Informative)
RISC does matter, or Intel and AMD wouldn't be using it.
Re:Technicly.. (Score:1)
Re:Technicly.. (Score:1)
True. Every computer interprets instructions, right the way down to the simplest level of switching a gate on or off.
The difference between RISC and CISC is the instruction set exposed to the programmer (compiler, interpreter, whatever), and thus the number of memory accesses needed to implement something complex. But if you were to believe Apple advertising (which anyone who actually buys Apples must do...), you would think less instructions is better.
I was simply demonstrating that anything Apple tell their lusers is believed, whether it's true, fabricated, or just hiding certain facts (like the relative floating point performance of the G4).
Feel free to disagree.
Re:If I had a supercomputer... (Score:1)
Air strikes against computers? (Score:2, Interesting)
Then, the US gets tired of bombing, and HP sells them new machines. Soon thereafter, we decide their new "good" dictator is just as bad as their old "bad" dictator, and the cycle begins again.
Re:Air strikes against computers? (Score:2)
Of course, the same could be argued about a Win2K or WinXP hacked clone - but the utility in solving nuclear equations and modeling explosions is not as high.
Even a suitcase bomb requires that you have:
1. component parts - detonator, nuclear material, shaping material
2. supercomputer to model the charge shape and impact velocities
3. willingness to deliver the material (even if lead shielded)
We know that they have #3, this HP open source supercomputer may give them #2, now they only have to pick up #1 - maybe Pakistan or the Taliban have such and will sell them to raise cash or create more problems for us.
Re:Air strikes against computers? (Score:2)
This might be different in for a country like Iraq who already have many educated physicists, and they have a realistic chance of actually doing all this work from scratch. Of course, the IEEE is doing random inspections there all the time, but maybe they could "disguise" their number-crunching supercomputer as a 256 separate workstation terminals for all the government clerks who write email. By night, thought, it's Linuxtime.
You're right about the missing material, but I'm sure someone somewhere will be willing to sell 10kg of it... (the more people we bomb, the more likely that seems).
spork
Re:The amusing implication of this (Score:2)
Re:Q3 Hmmm (Score:1)
When will people realize they can't just install Windows on a beowulf cluster ?
graspee