500 Billion Very Specialized FLOPs 89
sheckard writes: "ABC News is reporting about the world's fastest 'supercomputer,' but the catch is that it doesn't do much by itself. The GRAPE 6 supercomputer computes gravitational force, but needs to be hooked up to a normal PC. The PC does the accounting work, while the GRAPE 6 does the crunching." The giant pendulum of full-steam-ahead specialization vs. all-purpose flexibility knocks down another one of those tiny red pins ...
Crypto apps (Score:1)
Then you could say bye-bye to rc5-64. Perhaps before long you could eat rc5-64s like popcorn and go on to the other challenges at RSA [rsasecurity.com].
Like ENIAC (Score:1)
I guess this one is a little faster tough...
Gravity simulation algorithms (Score:5)
So along come some doods who said why don't we recursively stick the particles into boxes and then calculate the attraction between the boxes instead and it should be a lot faster. So they tried it and it seemed to work great- it only takes more like 10,000 calculations to do 1000 particles.
Anyway along came some other guys and they were a bit suspicious. They showed that some galaxies fell apart under some conditions with the recursive boxes method, when like they shouldn't. Back to the drawing board.
There are some fixes for this now- they run more slowly, but still a lot faster than the boring way. Still, its better than the end of the universe. Even if it is only a toy universe.
For descriptions of loadsa algorithms, including 'symplectics' which are able to predict the future of the solar system to 1 part in 10^13 ten million years in the future check out this link: [napier.ac.uk]
Imagine... (Score:1)
(sorry, i had to.)
Hmph... while some people worry that it is single purpose, they miss the fun... these people made a really fast computer. That's cool, by itself. It was created at the University of Tokyo, so it is obviously research, and not done as a cost-effective solution. I'm sure people can take lessons learned from this machine, and eventually apply it to a more broader market.
And having it controlled by a PC is no stranger then having your accelerated video card controlled by your computer, and it just doing the 3D video calculations. =^)
-legolas
i've looked at love from both sides now. from win and lose, and still somehow...
Re:What's the latest definition of "supercomputer" (Score:1)
and comes bundled with two onsite engineers
should be called a "supercomputer"
(the Jobs 'reality distortion field' G4 ads notwithstanding)
or perhaps anything that can crunch thru a
SETI [berkeley.edu] data block in 10 minutes!
MAB
Re:Imagine... (Score:1)
Re:Gravity simulation algorithms (Score:1)
As far as we know the 'gravity only' type of calculation that the Grape boards perform is sufficient to describe the motion of this matter.
However, there is indeed great interest in performing hydrodynamical simulations of galaxies, mostly because then we can attempt to calculate where and how the stars are forming in the galaxy. Dealing with the gas expicitly also allows us to follow things like shock fronts in the gas and to attempt to calculate the thermal properties of the gas. Of course this is all rather complicated stuff so we have to make gross approximations. And remember, even with that massive grape board if you describe a galaxy with a million particles, they are still going to each be representing at least 10 to 100 thousand solar masses. We are still a long way off from being able to describe the milky way on a 'star by star' basis.
Two really good URL's for people who are interested in reading some of the technical details of this stuff are the web pages of my advisor Matthias Steinmetz [arizona.edu] and one of the fathers of modern galaxy simulations Josh Barnes [hawaii.edu].
Note that Matthias's simulations (check out the pictures and movies) are all done with a high end workstation and a handfull of Grape 3 boards.
Cheers
Inaccurate! (Score:2)
tee hee
Re:"Specialised"? (Score:1)
The next two powers of 1000 are (Score:2)
Specialised v. General (Score:1)
A lot of dark matter could be big dust :-) (Score:2)
Nobody really knows how much of that stuff is out there. We know something is there that we don't see from the gravity puts out, but that doesn't mean it has to be something truly exotic.
Cheers,
Ben
Re:GRAPE-5 (Score:1)
What are these specialized "pipelines"? (Score:1)
Tera, not Peta. But Blue Gene is 1 Peta! (Score:1)
Gee I'm dumb. As several people pointed out, it's 100 Teraflops. Well, so much for my theory, "I don't get any dumber after a few beers."
As a consolation, here's a link to IBM's Blue Gene [ibm.com] supercomputer. It's still about 5 years off, but it will likely be the first Petaflop computer. It's being built specifically to solve a single problem--modelling protein folding. The best bit is that even at a petaflop, it will take about a YEAR [rutgers.edu] to simulate a single protein.
Re:Crypto apps (Score:2)
The Summer 2000 issue of American Heritage of Invention & Technology has a fascinating article on the specialized code breaking machines that were built and used during World War II.
I would tell, but government will kill me. (Score:2)
Search here [surfwax.com].
sounds like a rendring farm to me (Score:1)
There is also an other product that i cant remember the name of that acts like a rendering farm for 3D studio it has some custom rendering chips and an alpha for controlling it all. It actually runs linux...
Hey if we want to go on: the older multi processor Macs, had the second processor acting as a slave to the first one.
Im shore there are lots more examples, the story just made me think back on some cool rendering farm solutions that i have come a cross.
Re:bah! (Score:1)
Also, it sounds like you're arguing about ASICs vs. CPU's which is not what this is about at all. ASICs obviously are enormously useful (witness their vast dominance in the market), but it has nothing to do with whether or not you buy some custom supercomuter from SGI or build one yourself out of PCs and ethernet cabling for a fraction of the cost.
The ultimate special-purpose computer. (Score:1)
The planet is actually pretty expensive, but you can borrow it free of charge.
--
Do they use the theory of relativity? (Score:2)
Two words: Deep Crack. (Score:2)
embedded ram anyone? (Score:1)
playstation 2 has this and that is why it has a massive bandwidth of 48 gigs per second. Bitboys [bitboys.fi] has the same technology for the pc so lets hope they can actually release some thing.
I would like to know if any one is working on a processor whit embedded ram?
An other thing is the AGP bus that is just getting way to slow, and i guess that's up to Intel to do some thing about.
Re:bah! (Score:2)
Simple: various tasks need different amounts of bandwidth between the nodes to perform the calculation. For distributed.net and SETI@home, every data block is completely independent - the nodes don't need to communicate at all, so you just pipe the work units over the Internet.
Most problems don't break up this well, though - individual parts of the problem can interact with their neighbours, meaning individual nodes need to communicate with each other fairly quickly - a Beowulf cluster, for example. Lots of normal PCs on a fairly fast LAN.
Then, you have a handful of BIG number-crunching problems - like this one - where every part of the problem interacts with every other one. Think of it like a Rubik's cube: you can't just work one block at a time, you need to look at the whole object at once. This take serious bandwidth: the top-end SGI Origin 2800s run at something like 160 Gbyte/sec between nodes (in total).
Here in Cambridge, the Department of Applied Mathematics and Theoretical Physics has an SGI Origin 2000 series box with 64 CPUs - homepage here [cam.ac.uk]. (There's a photo of Stephen Hawking next to it somewhere on that site - this is his department.)
Basically, there are jobs clusters of PCs just can't handle. If the choice is between a $100k Beowulf cluster that can't do the job, and a $10m supercomputer which can, the latter is much better value.
Sure if you have the money to burn, go custom. But most of the computing projects out there do not require that kind of "big iron" and couldn't even afford it if they did. Besides, most of the time (unless you are in the DoD or NSA or such-like) you only end up with a small slice of that "big iron" which may or may not be roughly equivalent to being able to run your proggies on a computer that is all yours 24/7.
You're right - most projects don't need this kind of hardware. Some projects - including this one - do need it - either they cough up the big $$$, or the job doesn't get done.
Also, it sounds like you're arguing about ASICs vs. CPU's which is not what this is about at all. ASICs obviously are enormously useful (witness their vast dominance in the market), but it has nothing to do with whether or not you buy some custom supercomuter from SGI or build one yourself out of PCs and ethernet cabling for a fraction of the cost.
You can't build yourself a supercomputer out of PCs and Ethernet. You can build a cluster which will do almost all the jobs a supercomputer can - but not all of them. Some jobs need a supercomputer. A few very specialised jobs need even more muscle - like this one. It uses custom silicon, because that's the only way to get enough CPU horsepower.
Re:Reread the article.500 billion? Pah!100 Trillio (Score:1)
Do you not get it? This object only does one thing, can only do one thing, and is unable to do anything else. "Other calculations" are not possible because the algorithms are coded in silicone.
Re:The next two powers of 1000 are (Score:1)
Re:Gravity simulation algorithms (Score:1)
The recursive algorithm you described isn't the only particle-in-cell (PIC) game in town, incidentally. Perhaps the PIC techniques used in plasma simulations could be useful here? Plasma PIC simulations routinely model one or more conducting fluids with hundreds of millions of mutually interacting particles, often with comparable (in the case of electrostatic codes) or more complicated mutual interactions (in electromagnetic codes) than the blobs of gravitationally attracting fluids exhibit. (Instead of Newton's force law, in plasma media one solves Maxwell's equations to obtain the electric and magnetic fields, and then the particles are advanced in time using the Lorentz force). One thing that has resulted from this research is an understanding that in many parameter regimes of interest the "nearest-neighbor" interactions are less important than the collective effects, so smearing out individual particles into spatially extended blobs of superparticles can be a very reasonable approximation.
It's really sad seeing all these 'funny' posts (Score:3)
--
Re:"Specialised"? (Score:1)
Anyway, my point in all that was the the Cray's are designed for general purpose computation, even if they aren't designed to be as general as, say, database servers.
Grapes of Wrath (Score:2)
-------
CAIMLAS
Re:GRAPE-5 (Score:1)
$40K including an Alpha host and software. Only $10K for the actual superCruncher. Plus its small, so it shouldn't suck up that much power. This is much more powerful than a cluster of 5-7 linux pcs
Re:"Specialised"? (Score:2)
The only problem with previous versions of grape (that I know of) is that their precision is a little lower than you'd really like or need for some applications, but otherwise they are very nice for doing large n-body sims.
Doug
Yeah.... other Suns, for example (Score:2)
Yup. The Sun Enterprise 10000 [sun.com] (AKA "Starfire") uses a dedicated Ultra 5 [sun.com] as the console/management station. It connects via dedicated ethernet to the Starfire.
Re:Imagine... (Score:1)
http://www.nec.co.jp/engl ish/today/newsrel/0005/3001.html [nec.co.jp]
BTW, the NEC SX-5 - unlike massively parallel architectures - can effectively run near its theoretical peak performance for most applications, I'd say that the top 40 TFlops performance is a rather conservative estimate (NEC will have newer and faster technology by the time this beast starts being built).
The long-term definition of "supercomputer" (Score:2)
I always liked the definition, "Any computer that is worth more then you are."
;-)
many-body-system simulations (Score:1)
entropy. though most simulations suffer from reversibility (i.e. the system dynamics can be reversed: the simulated system evolves from state_x to state_y, but state_x can be determine exactly from state_y), researchers finally designed simulations that were not reversible (and the entropy correlated so well with theory you could derive boltzman's constant).
anyway, that's how i remember it, a passing comment from a class i really dug, but somewhere after debye shielding, i got lost--tensors can be rather difficult if you've spent most of your time writing code and designing circuits (hohlfeld, bless you, wherever you are ^_^;)
Re:"Specialised"? Not terribly novel, either (Score:1)
A lot of us who had been at TMC in the 1980's liked the CM-2 much more than the CM-5. Architecturally it was very clean. The CM-5 was a much more complicated machine.
Re:Special Problems. (Score:1)
The term is "symplectic integrator." You can check out the book "Dynamical Systems and Numerical Analysis" by A.M. Stuart and A.R. Humphries for an introduction and some references. The term refers to an ordinary differential equation solver that preserves the symplectic structure of the evolution semigroup of a Hamiltonian system. (Compare with Hamiltonian conserving methods). Such methods can be more accurate than general ODE solvers applied to a Hamiltonian system.
So, as far as I can tell, the poster made a typo but he isn't bullshitting. But you are probably a troll, so I'm not sure why I'm bothering.
Re:Reread the article.500 billion? Pah!100 Trillio (Score:1)
Ugh, (Score:1)
That's a lot of FLOPS (Score:1)
distributed.net (Score:1)
Re:"Specialised"? (Score:1)
"Specialised"? (Score:3)
Reread the article.500 billion? Pah!100 Trillion! (Score:4)
If you re-read the article, you'll see that 500 billion is just ONE OF THE BOARDS in the GRAPE. There are going to be 200 boards in this puppy, making for a machine that's getting 100 petaflops.
Damn fast!
bah! (Score:1)
The custom design ultra high performance on the order of a teraflops machines will still have their place at the top of the pile crunching stuff like quantum chromo-dynamics, simulated nuke blasts, and what-not, but the land of the middle of the line custom built number crunchers (from SGI, Sun, IBM, etc.) is quickly eroding.
Re:Do they use the theory of relativity? (Score:1)
The short scale problem can be resolved by various approximation methods, such as adding a softening distance term to the force calculation, and the long distance problem has pretty much been resolved by measurements (of the cosmic microwave background) that place the curvature of the universe at zero.
Even if this wasn't true these would still be good calculations to make, using pure Newtonian gravity, as it would allow differentiation between the behaviour of galaxies under Newtonian gravity and in the real universe.
I just wish I could have played with this baby for my computing project this year, which was the simulation of the collision of two galaxies (using very heavy approximations).
Re:That's 100 teraflops (Score:1)
___
Haiku (Score:5)
Processor of gravity
Quake sure feels real now
Re:bah! (Score:1)
When I spoke to David Ayd a few weeks ago, they were nearing the completion of a new generation , High Availability Public Access system based on a new supercomputer design meant for academic and research institutions. Unfortunately I can't release the details due to the NDA I'm under, but I can tell you that the project looks very exciting.
They wanted my advice on the implementation of the time-sharing protocol they're working on. It will literally make everything else look sick, the technology is awesome. Unfortuantely the protocol's design was beyond even me, and although I gave them some advice on it, it will take a few more years to complete the implementation.Cheers,
swImagine if they could do this for encryption. (Score:1)
Wonder where else can we use this? (Score:2)
That will help a lot...umm...while landing at Neptune some day.
Re:"Specialised"? (Score:2)
Tell a man that there are 400 Billion stars and he'll believe you
Re:bah! (Score:1)
GRAPE 4 (Score:1)
Tell a man that there are 400 Billion stars and he'll believe you
Almost. (Score:2)
There is a lot of work in Processor-in-memory (Score:4)
There are many problems with implementing a system like this in practice. The fabrication process used for DRAM's is completely different from that used for logic. In general, for DRAM you want a *high* capacitance process so that the wells holding your bits don't discharge very quickly -- that way you can refresh less often. In logic you want *low* capacitance so that your gates can switch quickly (high capacitance -> high RC time constant -> slow rise/fall time on gates -> slow clock speed).
Fabricating both with the same set of masks doesn't work particularly well, so you really have to compromise -- you'll basically be making a processor with a RAM process, or vice-versa. Alternately, you could use SRAM, which is nice and fast and is built with a logic process, but is 1/6th the storage density of DRAM. This is why SRAM is used for caches and DRAM is used for main memory.
Having the memory on the same die as the processor definately gives a bandwidth and latency advantage. For instance, when you are on the same die, you can essentially lay as many data lines as you like so that you can make your memory interface as wide as you like.
But another large advantage is the power-savings. Processors consume a great deal of their power in the buffers driving external signals. Basically, driving signals to external devices going through etch is power-expensive, and introduces capacitances that kill some of your speed. Keeping things on die, no such buffers are needed, and a great deal of power is saved.
The first commercial application of the processor-in-memory concept that I am aware of is Neomagic's video cards. They went with PIM not for bandwidth, but for power-conservation, and chip reduction. These characteristics are extremely appealing to portable computing, and thus Neomagic now pretty much owns the laptop market.
In a limited application, such as a 2D graphics card, this is feasible because the card only needs perhaps 4 MB of memory. Placing an entire workstation's main memory (say, 128 MB) on a single die *with* a processor would lead to a ridiculously massive die. Big dies are expensive, lead to low yield and increase design problems with clock skew. Thus, having 128 MB of DRAM slapped onto the same die as your 21264 isn't going to happen in the near future.
Placing a small (4-8 MB) amount of memory on-die, and leaving the rest external is possible, but leads to non-uniform access memory, which complicates software optimization and general performance tuning greatly. It is generally considered undesirable.
Another approach is to build systems around interconnected collections of little processors, each with modest computing power and a small amount (say 8 MB) of memory. Thus, you are essentially building a mini-cluster, where each node is a single chip. This, too, leads to a NUMA situation, but it is more interesting, and many people are pushing it.
PIM's are going to be used more and more, and the massive hunger for bandwidth in 3D-gaming cards very well may drive it to market acceptance. The power consumption adavantages will continue to appeal to portable and embedded markets as well. However, general purpose processors based on this design are unlikely in the near future. This style of design doesn't mesh well with current workstation-type architectures.
A bit of a tangent, but I hope it was informative...
--Lenny
The solution is trivial (Score:2)
The solution is trivial.
1. Carry Ultra-Sparc to building rooftop.
2. Drop Ultra-Sparc off building rooftop.
3. If results are disputed, request that critic stand at base of building. Repeat steps 1 & 2.
Re:Gravity simulation algorithms (Score:1)
Re:Like ENIAC (Score:1)
Somebody correct me if I'm wrong, but I'm pretty sure that the ENIAC was used for calculating artillery tables, not ballistic missile paths...
Re:GRAPE-5 (Score:1)
other problems expressed as gravity problems? (Score:1)
Step 1) Acquire data on the purchasing behavior and demographic info of a couple of million consumers from some unscrupulous web retail site.
Step 2) Get a few scaling variables on the front- and back-end, replace stellar mass with income, replace stellar velocity with purchasing habits, replace stallar cluster density with population density (or proximity to retail outlets), etc., etc.
Step 3) Run the system to model consumer purchasing decisions for a product you're planning to introduce into the marketplace.
Surveys measure economic activity on a large scale and make broad predictions. Could this be used to more accurately model and predict economic behavior on a more precise scale? The data would be constantly updated, and the models would be constantly rerun to get the most accurate picture possible of how you and I will spend our $$$. Just make sure the the observed isn't aware of the observation, or your models lose their viability.
What these boards really are... (Score:2)
Aww, talk about sour grapes! They've hurt IBM's feelings, because IBM sells really smokin' computers too.
Seriously, I think David misclassified GRAPE 6 quite a bit. I don't think it's quite David's fault, because the article writers don't know the difference between 'supercomputer' and 'attached processor'. ABC News didn't really apply the term 'supercomputer' correctly either.
The term 'supercomputer' is more of a marketing term than anything else. Technical people only use it when they want to describe a general capability. AFAIK there is no concrete definitions of 'supercomputer', and if there were they would likely change daily. GRAPE 6, from the information I can see, is really an attached processor.
Attached processors can be an ARM chip on your network card [3com.com] to a GRAPE 6. Interanally, GRAPE 6 is a full custom, superscalar, massively pipelined, systolic array (say that 5 times fast). That basically means that data comes in one side of the board, and after n clock cycles the answer comes out the other side. There is no code other than a program running on the host computer which generates and consumes data, and every piece of the algorithm is done in hardware.
"What happens when the algorithm changes?" you might ask. Well, then you're screwed. You have to do a whole new board. Many boards use programmable chips as their processing elements, and can reprogram them when bugs or features get added, but these guys appear to be using ASICs. Great for speed, bad for flexibility.
Even though David Ayd was mistaken about the architecture, this idea has been around for quite a while also. The SPLASH 2 [ccic.gov] project was one of the first successes with this idea. There is also a commercial company [annapmicro.com] selling boards using that idea but with completely up to date components (compared to SPLASH).
Still, in July of 1995, the GRAPE 4 became the world's fastest computer, breaking the 1 teraflop barrier with a peak speed of 1.08 TFLOPS.
Well, we really can't argue with that, can we, Mr. Ayd?
This architecture lends itself to extremely high throughput. It's no surprise that these perform so well. NSA uses architectures just like this to do it's crypto crunching. Brute forcing doesn't look so bad after trying one of these
top tier of performance & price (Score:1)
Considering a Mac G4 chip peaks at 4 GFLOPS
Re:Gravity simulation algorithms (Score:1)
I think most of the infomation is still fairly accurate. The paper is aimed at semi-techincal people, but not experts in the field of the n-body problem. That is, show some formulas that such that use basic, first year calculus, but you don't need to know how to use a "Hamiltonian Operator".
You can find a copy of XStar program and paper can be found here [midwestcs.com].
The original document was created in FrameMaker and I have been unable to fully HTML-ize it. To get a copy with pictures, you need to look at the postscript document.
supercomputer is oxymoron (Score:1)
In exchange for increased performance in some
repect, you lose something in general purpose
computing, such as software tools, programming
generality, adequate peripherals etc.
Re:A lot of dark matter could be big dust :-) (Score:1)
Actually there are pretty reasonable arguements against that possibility. The strongest of which is cosmological. If you believe cosmology the relative abundance of the light elements (hydrogen, helium, and lithium) would be thrown all out of whack if the universe has more baryons (stars, dust, moons, gas and such) than about a third of the mass needed to make the universe flat. Basically this has to do with the fact that during the first few moments (of the universe) nuclear reactions are going on all of the time turning hydrogen into helium and back again. The forward and backwards reactions are density dependent and go on for as long as the universe is hot enough to sustain the reaction. So the relative abundances give you a measure of the density of the baryons in the universe at the moment that the universe cooled enough to stop the reaction.
The universe appears to be flat, from redshift surveys and the ripples in the cosmic microwave background. So since we know that baryons can't do more than a third of that we are forced to postulate something wierd to account for at least the other two thirds.
The most likely value of the baryon fraction is actually around 12 percent and the rest is split between dark matter and something even wierder called 'vacuume energy of space' (or 'lambda').
Hope that made some sense.
chris
it's your universe, get used to it
Underclocking (Score:1)
Re:"Specialised"? (Score:1)
Re:"Specialised"? (Score:1)
Re:Reread the article.500 billion? Pah!100 Trillio (Score:2)
Secondly, that's "theoretical peak performance", otherwise known as the "guaranteed not to exceed" performance. On their highly specialized code it'll probably do ok, but on other calculations I'd be surprised if it got 10% of that speed, especially if a lot of cross-node communication is occuring. Don't forget, this is not a general purpose computer, it's like a really really big math co-processor that is optimized to run a very very specific type of program fairly well.
Irony? Perhaps. (Score:1)
Computing trajectories.
(Disclaimer: Yes, I know it's only one of the cutting edges, and yes, I know gravitational interactions aren't strictly the same as trajectories, but the irony remains, okay?)
Re:Reread the article.500 billion? Pah!100 Trillio (Score:1)
I need this NOW!!!!!! (Score:4)
How hot does it run? (Score:1)
GRAPE-5 (Score:3)
What's the latest definition of "supercomputer"??? (Score:2)
Which reminds me, if anyone is interested in the "flopsability," to coin a silly-sounding word, of common x86 processors, visit http://www.jc-news.com/parse.cgi?pc/temp/TW/linpa
Re:distributed.net (Score:1)
oh yes! (Score:1)
Re:Haiku (Score:1)
Temperature far too high
Expensive doorstop
Re:bah! (Score:1)
argh ! troll ! (Score:1)
Stop trolling! The Steve Woston [mnc.co.za] is terribly annoyed at being impersonated by trolls on /. Read what the Real Steve Woston has to say about it here. [mnc.co.za]
Special Problems. (Score:3)
I have a cluster of alphas crunching away solar system models - Grape6 couldn't actually do this very well since it's designed for a certain N body algorithm which doesn't suit small N... Instead I use a syplectic integrator which takes advantage of a number of known factors in the problem.
So - we still need bigger and faster machines, but we also need more general machines...
Anyway... I want one of these to model EKO formation in the solar system
Re:GRAPE-5 (Score:1)
Re:bah! (Score:2)
"General"... (Score:1)
-jhp
Re:What's the latest definition of "supercomputer" (Score:4)
The problem with anything based on a microprocessor is the pathetic main memory bandwidth. If your program blows out the cache, the performance goes to hell.
A vector supercomputer is designed to have massive memory bandwidth, enough to keep the vector processing units operating at high efficiency. No cache or VM to slow things down. An engineer once told me that a Cray was a multimillion dollar memory system with a CPU bolted on the side.
See the STREAM benchmark [virginia.edu] web page for some measurements of sustained memory bandwidth. This separates the real computers from the toys.
Quick-to-diss IBM (Score:2)
Aww, talk about sour grapes! They've hurt IBM's feelings, because IBM sells really smokin' computers [ibm.com] too. But:
Still, in July of 1995, the GRAPE 4 became the world's fastest computer, breaking the 1 teraflop barrier with a peak speed of 1.08 TFLOPS.
Well, we really can't argue with that, can we, Mr. Ayd?
--
"Give him head?"
"One World, one Web, one Program" - Microsoft Ad
FIR Filters and Neural Networks (Score:3)
I haven't followed the progress in the field since then, but I suspect present day hardware could handle a good fraction of the satellite image feeds affordably -- and dwarf the realized performance figures of this gravitation board.
Of course, if you want to get really picky about it, there are lots of specialized circuits out there doing work all the time all over the place that could be viewed as "computation" at enormous rates -- it all depends on where you draw the line.