TeraGrid v. Distributed Computing 124
Nevyan writes "After three years of development and nearly a hundred million dollars the TeraGrid has been running at or above most peoples expectations for such a daunting project. On January 23, 2004 the system came online and provided 4.5 teraflops of computing power to scientists across the country. However, the waiting list for TeraGrid is long, including a bidding process through the National Science Foundations (NSF's) Partnerships for Advanced Computational Infrastructure (PACI) and many scientists with little funding but bright ideas are being left behind. While the list of supercomputer sites and peak power is growing how is the world of Distributed Computing faring? "
Distributed Computing (Score:5, Insightful)
Distributed computing has its uses, but remeber: the public will only be willing to help you as long as they feel like they're contributing to something worthwhile.
Re:Distributed Computing (Score:4, Insightful)
I think that this is a Good Thing (TM). Distributed computing has the postential to not only further the cause of science, but to bridge the gap between the public and the scientists.
Re:Distributed Computing (Score:3, Interesting)
Only if they are investigating cannibalism. The purpose of science is the advancement of knowledge. Service to humanity, if it happens is incidental.
Re:Distributed Computing (Score:3, Insightful)
Re:Distributed Computing (Score:5, Insightful)
What I'm trying to say is that the semantics of how we describe the universe may be arbitrary, but the universe is objectively describable.
Re:Distributed Computing (Score:1, Insightful)
Science doesn't have purpose. It's a study of phenomena. At least, that's the scientific method:
Re:Distributed Computing (Score:2)
Science teaching not at fault- make English easier (Score:2)
To list only a few rants:
Re:Science teaching not at fault- make English eas (Score:2)
Who would have guessed that ignorant hicks could improve the English language?
I suppose it makes sense, though, that someone with less education could come up with simple & obvious solutions (although I'd prefer using a word that didn't make you sound stupid).
Re:Distributed Computing (Score:5, Insightful)
Jesus, there's a horrible thought. I've met the public (and seen it's choice in TV). I'd rather have monkeys choose.
Why being so negative? (Score:2)
Now the public can chose what problems that it wants solved
Jesus, there's a horrible thought. I've met the public (and seen it's choice in TV). I'd rather have monkeys choose.
Well the idea is that, you, as a member of said public, take responsibility more serious instead of just dissing it because others do.
Your Gorilla agent (Score:1, Interesting)
You might be right about those monkeys. In Holland, we have the Beursgorilla (http://www.beursgorilla.nl/ [beursgorilla.nl]). This gorilla decides what stock to buy or sell based on the bananas presented to him. He proves to be better at "advising" than most of the other "real" and expensive advisors.
For me, the DistributedComputingGorilla might decide what project will run on my computer.
Re:Distributed Computing (Score:3, Insightful)
Re:Distributed Computing (Score:2)
I do not see how this is in anyway bridging the gap between the public and scientists. How is donating your free computer cycles any different than donating a few bucks? You are just donateing a resource cash or cycles to a project.
I mean it is not like most people running Seti at home know what an FFT is.
Not a bad thing min
Its a Cookbook (Score:2)
Disgusting. Your Nickname says it all. My only hope is that after I'm long dead, you
Meta-programs for distributed computing (Score:3, Informative)
Re:Distributed Computing (Score:5, Interesting)
Personally, I'd much rather have an applet using 10% of my cpu power instead of an annoying flash banner (which probly itself uses 10% cpu...).
Obviously someone has to pay for internet content, and that to me would be the least intrusive way. Popup-blockers will be inefficient by the end of the year. Ads will be inside the site content. Or worse still, the popup window is the main window, while the actual content is spawned as "pop under" meaning that if you have a popup stopper, all you get is the ad window...
Cpu cycles is the perfect internet currency. Everyone who visits a website has them.
Re:Distributed Computing (Score:2, Interesting)
Re:Distributed Computing (Score:3, Informative)
Off-topic. Teragrid is a dedicated distributed computing system. Various research centers are purchasing dedicated clusters to participate. For example, instead of three universities each purchasing a large cluster which will sometimes be idle; each will purchase a slightly smaller cluster and use each other's resources when available. In
My Personal Vision (Score:3, Interesting)
Re:My Personal Vision (Score:5, Informative)
Re:My Personal Vision (Score:5, Informative)
Re:My Personal Vision (Score:1)
Current and Future BOINC projects (Score:2)
Currently available projects are...
SETI@home [berkeley.edu]
Predictor [scripps.edu]-Protein structure prediction
Coming soon....
climateprediction.net [climateprediction.net]
Folding@home [stanford.edu]
Farther in the future (i.e. pending funding)...
Einstein@home [physics2005.org] -- a search for gravitational waves.
In the conceptual stage, since sometime last week...
neuralnet.net -- studies of the nature of intelligence using neural nets and genetic algorithms
Payment for Work Units (Score:5, Interesting)
1) Getting the money to pay people. One advantage of distributed computing is that you don't have to pay for time on expensive cluster. That advantage disappears when you pay distributed computing users. Of course, it may still turn out to be cheaper, and there may be users willing to participate for free.
2) Botnets and profit. We all know of spammers using zombies to peddle goods, and of script kiddies using them to DDoS. What if some enterprising but immoral person decided to use the computing power of his zombies to profit off of the distributed computing payments? With enough zombies, he could easily make a good amount of money off of other people's computers.
Re:Payment for Work Units (Score:5, Interesting)
Re:Payment for Work Units (Score:2, Interesting)
Re:Payment for Work Units (Score:2)
Re:Payment for Work Units (Score:2, Insightful)
2) Yes, security is a big issue in Grid computing. And it ain't there yet.
Re:Payment for Work Units (Score:2)
So? I'd rather the spammers (or "clustered computer users" in this case) used their zombie machines for that. Then it would get the spam to stop.
Re:Payment for Work Units (Score:2)
The only bad possibility is that it might increase the number of zombies. However, this is not necessarily so. It is not evident to me that it would be simple to incre
Re:My Personal Vision (Score:5, Interesting)
They are in beta stages of a massive computation cycle for hire program that will allow organizations without the funding for an entire cluster to purchase cycles provided by a large IBM Power cluster.
It will allow for a computation cycle market to eventually arise, much like the wheat, corn or gold markets. Companies will compete to provide cheaper cycles, small-time scientists around the world will be able to have thier computation intensive problems solved at a fraction of the current cost possible today.
Re:My Personal Vision (Score:2)
Re:My Personal Vision (Score:2)
Re:My Personal Vision (Score:2)
Re:My Personal Vision (Score:2)
I'll bet there tons of people that would donate processing time to helping defend open source.
Re:My Personal Vision (Score:2)
I wouldn't mind it if I could make a little cash to eventually help pay back for the computer, maintainance and energy (and heat removal in the summer), I doubt it would happen. I'd love to see someone prove me wrong.
Looks good to me (Score:4, Interesting)
There are some like the Casino-21 http://www.climate-dynamics.rl.ac.uk/ [rl.ac.uk] and Evolution-at-Home http://www.evolutionary-research.org/ [evolutiona...search.org] too.
It's becoming easier to create the required code for distributed projects, and it most certanly has become easier to actaully get them distributed.
Cue the jokes... (Score:2, Funny)
But will it run Longhorn?
Re:Cue the jokes... (Score:3, Funny)
Is Apple giving away an iPod to whoever uses this joke for the 100,000,000th time or something?
Re:Cue the jokes... (Score:1)
Grid and Distributed comptuing (Score:5, Informative)
What
Re:Grid and Distributed comptuing (Score:1, Interesting)
There are various collections of machines which have been designed to facilitate Grid computing (instances of Grids), TetraGrid being one of them. Some systems or Grids are suitable for some types of jobs, some for others. As you rightly note, for the likes of MPI you need relatively closely coupled nodes.
Re:Grid and Distributed comptuing (Score:1, Interesting)
Grid" or Grids is like "The Internet" vs multiple
IP-protocol networks, including the private ones.
However, for practical purposes there is one "The
Grid" which will probably evolve into The Grid
without the quotes, and that is the worldwide
LHC Computing Grid, currently spread across North
America, Europe and northwest Pacicifc Rim.
Through EGEE (in Europe) and Open Science Grid
(in the US) LCG technology will spread out into
the wider scientific and research commun
The Google Compute Project (Score:5, Informative)
Google's distributed OS has been discussed a lot on Slashdot, but it is more than just a search algorithm on their own servers:
Google Compute is a feature of the Google Toolbar that enables your computer to help solve challenging scientific problems when it would otherwise be idle. When you enable Google Compute, your computer will download a small piece of a large research project and perform calculations on it that will then be included with the calculations performed by thousands of other computers doing the same thing. This process is known as distributed computing.
The first beneficiary of this effort is Folding@home, a non-profit academic research project at Stanford University that is trying to understand the structure of proteins so they can develop better treatments for a number of illnesses. In the future Google Compute may allow you to also donate your computing time to other carefully selected worthwhile endeavors, including projects to improve Google and its services.
- The Google Compute Project [google.com]
Re:The Google Compute Project (Score:1)
already done! (Score:3, Funny)
Just think, for once redmond got it right! Windows has had that feature for years!
Re:The Google Compute Project (Score:4, Informative)
The distributed OS and Filesystem in thier own clusters is far more advanced than a SETI@Home parallel work distribution algorithm. This OS/FS and projects like it are where the grid's heritige lies. There are many problems unique to the grid, but none of it could exist without the distributed system problems first solved in local area clusters.
TeraGrid doesn't use "Public" computers (Score:3, Informative)
Re:TeraGrid doesn't use "Public" computers (Score:2)
Re:TeraGrid doesn't use "Public" computers (Score:1)
> Uh, I'm not sure what this has to do with the TeraGrid . . . The TeraGrid is a distributed computing system . . . but it does not use the "public's" computers.
Oh, the header was "TeraGrid v. Distributed Computing" and the entry ended in a phrase "While the list of supercomputer sites and peak power is growing how is the world o
Re:TeraGrid doesn't use "Public" computers (Score:2)
They are public in the sense that they are publically funded. They are not owned by a private corporation or restricted to classified use.
Re:TeraGrid doesn't use "Public" computers (Score:2)
Re:TeraGrid doesn't use "Public" computers (Score:2)
Correct. Usage is not restricted to the people who work there. In fact, 99.999% of the work done on the NSF supercomputing resources is done by people who do not work at the supercomputing centers. Researchers from U.S. institutions apply for time on the supercomputers. The centers are a computing resource for the whole country.
Re:TeraGrid doesn't use "Public" computers (Score:3, Informative)
Despite that fact, you are correct... most of the work that's run on those things is done by people that aren't part of the supercomputer centers, or the ANL. (There are a few "chief scientists" that DO run their work there, so I wouldn't say it's the 99.99-whatever% that another poster did).
It's NOT available for the general public's use though. Even if you work at those places, that doesn't give you ANY certaint
Re:TeraGrid doesn't use "Public" computers (Score:2)
Thanks, you said what I was thinking better than I was able to. I don't care who funds it, very few people are going to call something "public" if the general public cannot access it.
Just imagine a Beowulf cluster of... (Score:2, Funny)
Look, this is really very simple (Score:5, Insightful)
So: SETI@Home splits up its scans into sections, each of which do not depend on any other; therefore, a distributed solution is efficient. However, the Earth Simulator deals with chaotic systems (or so I would assume), which do not independently parallelize; this is where having hundreds of processors and terabytes of RAM and using something like NUMA is greatly more efficient.
In short: use the right tool for the job.
Re:Look, this is really very simple (Score:5, Informative)
Re:Look, this is really very simple (Score:1)
AFAU Earth Simulator solves mostly nothing more than a big Finite Element Method problems. Speed-up of such problems depends much on the connection time as normally the FEM solver exchange borders every several iterations or so, while the amount of data is not so much
Access and Denial (Score:4, Insightful)
While TeraGrid is a powerfool tool it is one that thousands of scientists and laboratories are standing in line to use. Meanwhile Distributed Computing is available, cheap and relatively quick.
While it may look good on your project to say you used a IBM BlueGENE or DeepComp 6800 is it really worth the extra cost and waiting in line for your chance to use?
True Distributed Computing is the way to go and shows positive results. Now we just need to tinker with it some more!
Re:Access and Denial (Score:2)
Re:Access and Denial (Score:5, Interesting)
What are you talking about? These are publically funded resources. You apply to the NSF for time on these machines. If you're at a U.S. institution and you have a real need for supercomputing you can get time on these machines.
And Distributed Computing can't even begin to solve some of the problems that supercomputers are designed to address.
Yes. When you want to simulate every molecule of a proteing in a water solution (~17000 atoms worth) you need a supercomputer. DC can't do it.
DC is neither a religion nor a panacea.
Re:Access and Denial (Score:3, Insightful)
> proteing in a water solution (~17000 atoms worth)
> you need a supercomputer.
But if you want to simulate a billion molecules,
DC is the way to go: Then it's not a tightly
coupled system.
Re:Access and Denial (Score:2)
True, if you're not simulating how those molecules interact. Don't get me wrong, DC is great for a lot of things. My point is that for certain problems DC just isn't up to the task.
Re:Access and Denial (Score:2)
Re:Access and Denial (Score:2)
Most applications will face a performance plateau at
Most scientific applications are network intensive, and 2 CPUs with a local, high-bandwidth/low-latency connection (GigE, IB, Quadrics) could very well outperform 100 CPUs distributed across the internet.
Only for very, ver
Re:Access and Denial (Score:2)
The TeraGrid is distributed-computing but not like what the poster is calling DC. The poster is talking about things like SETI and encryption cracking. They are massively parrallel -- i.e. problems that can be broken up up into discrete units and worked on by unconnected machines.
Anyway, I'm for both and against misinformation about what both are. The poster confuses the issue terribly.
Re:Access and Denial (Score:2)
Suppose you're not at a U.S. institution and/or have such
Re:Access and Denial (Score:2)
There is no conspiracy. Supercomputers solve problems that DC cannot begin to address. Your example DC simualtion wouldn't work because of the time it would take to communicate all that information. The latency issue alone would make it unsusable.
Do you really think only prestigous scientists get to use supercomputers? That the little guy is being intentionally kept down by the supercomputing man? The idea is ludicrous. If someone has a truly 'brilliant idea' they will get on the superomputer.
There are
Re:Access and Denial (Score:2)
But I'm not, so your logic is wrong.
I use supercomputers all the time.
I call BS (Score:5, Informative)
It's too bad that whoever modded this Insightful doesn't know much about parallel applications.
DC is fine and very cost-effective for its niche of applications, which is those that are "embarassingly parallel." This is (somewhat circularly) defined as being very easy to parallelize on a DC machine. What characterizes these apps is very low communications between different tasks, which works for DC because the high network latency doesn't get in the way.
I've love to see you try to put Conjugate Gradiant (CG) on a distributed system. It involves large matrix-vector multiplies that inherently require lots of vector fragments passing between the processors. CG is one of the 8 NAS Parallel Benchmarks, and if you look at Beowulf papers that use NAS, you'll see that they often leave out CG because performance is so bad. If it's low on a Beowulf, where the network is presumed to be local and dedicated, it will totally suck on anything with a typical high-latency/low-bandwidth network.
Re:Access and Denial (Score:1)
They should go elsewhere. If they're really scientists they probably have a full time job at an institution with some kind of cluster or are affiliated with an institution that can provide access to resources commesurate with the scientist's ability.
If there's a guy with no previous record who claims he can create AI if they let him use TeraGrid exclusively for 6 months, should he be given access to TeraGrid
Why the versus? (Score:5, Insightful)
Re:Why the versus? (Score:1)
We're not, we're asking how TeraGrid is doing compared to Distrubuted Computing. Read the Fine Title!
What? Oh? Nevermind...
Did I read that right? (Score:4, Insightful)
Re:Did I read that right? (Score:2, Informative)
Pittsburgh Suprecomputing Center joined in when NSF announced supplementary funding with $35 million.
$10 million was supplied by NSF in September 2003 adding ORNL, Purdue, Indiana U., and TACC.
Total: $98,000,000.00 roughly.
What does government spending on the TeraGrid give you? 4.5 Teraflops distributed...
Nice.
Re:Did I read that right? (Score:2, Informative)
Yes, the money was spent somewhere, and not on toilet seats.
Re:Did I read that right? (Score:4, Informative)
I would encourage you to visit www.teragrid.org [teragrid.org] and read more than the front page to learn what the Teragrid is
Re:Did I read that right? (Score:5, Insightful)
Re:Did I read that right? (Score:4, Informative)
One project [psc.edu] on the Teragrid used 17 Teraflops. The poster hasn't done his/her research on the Teragrid.
Mac OS X users! (Score:2, Informative)
Misinformed (Score:4, Informative)
Care to cite a source?
When you apply to the PACI program you get a grant of Service Units -- i.e. time on the computers. You don't need huge amounts of funding. The requirements [paci.org] state that you need to be a researcher at a U.S. institution. It also helps if you can show that you actually need and can use that kind of computing power.
And, please, distributed computing and supercomputing are not synonymous in terms of what problems they address. Distributed computing cannot replace supercomputers in every case. DC is good for a limited set of problems.
Lastly, an example of Teragrid research: Ketchup on the Grid with Joysticks [psc.edu].
Wolfgrid (Score:5, Informative)
It is based on Apple's XGrid, and uses volunteers from the Mac community here at NCSU, as well as some of the lab macs, and soon we will hopefully have official Linux and Windows clients, maybe even Solaris, to run on more of the computers around campus.
There is even a really nice web interface [ncsu.edu] that shows the active nodes and their status, as well as the aggregate power of the two clusters.
Its really nice, anyone who is part of the grid can just fire up the controller and submit a job, I am a part of the lower power grid since my TiBook is only a 667, but I was able to connect up and do the Mandelbrot Set thing that comes with XGrid at a level equal to around 7 or 8 GHz.
There are some screenshots here [ncsu.edu]
Recommend good cause to donate my free cycles to? (Score:2)
-Brien
Re:Recommend good cause to donate my free cycles t (Score:4, Informative)
http://www.stanford.edu/group/pandegroup/folding/ [stanford.edu]
But I'm quite selfish (and actually interested in primes abd or at least know more about them than I do about protiens), and there are entities offering big prize money for big primes, and if one of my machines finds one, I'll get big bucks:
http://mersenne.org [mersenne.org]
Re:Recommend good cause to donate my free cycles t (Score:4, Interesting)
While there's much to learn from studies of protein folding, there's very little medical importance to purely theoretical simulations. Since the delusion that we'll be able to replace laboratory research with really big computers is attractive to people who know nothing about biology, the impact of this type of research gets vastly overstated.
On the other hand, Folding@Home has already yielded far more interesting results (if not exactly "useful" outside of the world of biophysics) than SETI@Home probably ever will, so go for it.
Re:Recommend good cause to donate my free cycles t (Score:3, Insightful)
Re:Recommend good cause to donate my free cycles t (Score:1)
The key word there is "probably." One positive, one single positive, from SETI will be arguably the greatest discovery in human history (certainly in the top 5).
Re:Recommend good cause to donate my free cycles t (Score:1)
Re:Recommend good cause to donate my free cycles t (Score:1)
yes I know my sig is coming up
Re:Recommend good cause to donate my free cycles t (Score:2, Informative)
This isn't Grid Computing. (Score:3, Informative)
It looks like TeraGrid is latching onto a catchword in order to boost awareness of their system. What they are describing here is not Grid computing at all. Grid computing was designed to take advantage of all the dead cycles that computers typically have. The idea is that someone might have a large group of computers that do not take full advantage fo their computational cycles (like a large lab for reading e-mails and browsing the Internet). With Grid computing you would take these computers (not some Itanium cluster like TeraGrid is doing) and distribute work accross these nodes that can be performed during otherwise dead cycles. (I have no sources immedeately available but check out Grid computing through the ACM or something and you'll see plenty of info on what Grid computing really is.)
This is what Seti@home does. It takes underutilized machines and runs computations on them. TeraGrid on the other hand, takes large clusters of otherwise unused machines and lays an abstraction over them that makes them look like one large supercomputer. This is nothing more than a distribution strategy. It looks like a nice distribution system that has the potential to scale well, but it's not Grid computing and it's nothing new.
Re:This isn't Grid Computing. (Score:1, Interesting)
It looks like TeraGrid is latching onto a catchword in order to boost awareness of their system. What they are describing here is not Grid computing at all."
No, they are right and you are wrong.
Using spare cycles is one thing you can
do with Grid technology, but it is not the
essential qual
Re:This isn't Grid Computing. (Score:3, Informative)
The TeraGrid people (ANL, Ian Foster, etc) are the ones that coined the term "Grid" in the first place!
You might not like their use of that term, but since they're the ones that came up with it in the first place, they're more right than you are.
Teragrid runs a lot of Linux (Score:4, Informative)
is AIX 5.2, 3128 (WOW!!) is on Tru64 (in 2 clusters) and the rest, distributed in 5 clusters
are some form of Linux.
Two of the clusters have a second phase which together will add 316 CPUs on Linux.
As of October 1 of this year, 5 clusters at 3 sites will be added with the OS / CPU breakdown as follows:
Linux : 1800 CPUs in 3 clusters
AIX 5.1 : 320 in 1 cluster
Solaris 9 : 256 in 1 cluster
That's an awful lot of Unix and a buttload of Tru64 and Linux