Become a fan of Slashdot on Facebook

 



Forgot your password?
typodupeerror
×
Supercomputing Technology

The Supercomputer Race 158

CWmike writes "Every June and November a new list of the world's fastest supercomputers is revealed. The latest Top 500 list marked the scaling of computing's Mount Everest — the petaflops barrier. IBM's 'Roadrunner' topped the list, burning up the bytes at 1.026 petaflops. A computer to die for if you are a supercomputer user for whom no machine ever seems fast enough? Maybe not, says Richard Loft, director of supercomputing research at the National Center for Atmospheric Research in Boulder, Colo. The Top 500 list is only useful in telling you the absolute upper bound of the capabilities of the computers ... It's not useful in terms of telling you their utility in real scientific calculations. The problem with the rankings: a decades-old benchmark called Linpack, which is Fortran code that measures the speed of processors on floating-point math operations. One possible fix: Invoking specialization. Loft says of petaflops, peak performance, benchmark results, positions on a list — 'it's a little shell game that everybody plays. ... All we care about is the number of years of climate we can simulate in one day of wall-clock computer time. That tells you what kinds of experiments you can do.' State-of-the-art systems today can simulate about five years per day of computer time, he says, but some climatologists yearn to simulate 100 years in a day."
This discussion has been archived. No new comments can be posted.

The Supercomputer Race

Comments Filter:
  • by Anonymous Coward on Tuesday September 23, 2008 @06:49PM (#25129355)

    Is how many libraries of congress it can read in a fortnight.

  • Simulation (Score:5, Funny)

    by gringer ( 252588 ) on Tuesday September 23, 2008 @06:55PM (#25129393)

    Simulate 100 years of climate in a day? Here's my code:

    echo -e "sunny\nrainy\ncloudy" | rl -rc 36525

    • What is rl?
      • Re: (Score:3, Funny)

        by Kneo24 ( 688412 )
        Relive! The machines have sentience! RUUUUUUN!
      • Re:Simulation (Score:4, Informative)

        by monsul ( 1342167 ) on Tuesday September 23, 2008 @07:17PM (#25129609) Homepage
        http://ch.tudelft.nl/~arthur/rl/ [tudelft.nl]
        From the webpage: rl is a command-line tool that reads lines from an input file or stdin, randomizes the lines and outputs a specified number of lines. It does this with only a single pass over the input while trying to use as little memory as possible.
        Didn't know about it either. Seems marginally useful
      • by Surt ( 22457 )

        The Random Line picker.

      • by gringer ( 252588 )

        It's a simple line randomisation program. I use it a lot in the work that I do for short shell scripts, which involves doing lots of permutations of lines of text files (or just sequences of numbers). Once that code gets put deeper into a loop (and hence becomes more of a limiting factor in terms of excecution time), I substitute it for faster stuff written in Java, Perl, R, C++, or whatever else takes my fancy at the time. shuf is a similar program which seems to be in the coreutils of Debian, but doesn't

    • Re: (Score:3, Interesting)

      by GroeFaZ ( 850443 )
      You jest, but that's exactly the point. "Simulated years per day" is about as meaningless a metric as it gets, because, as you proved, that number depends on the complexity of the underlying climate model, and also on how well the software was written, i.e. if it is optimized for both the hardware and the model to be computed.

      Both these factors are hard/impossible to control and to standardize, and the only factor that does not change is the actual hardware and its peak/sustained performance, so it's the o
  • by Ritz_Just_Ritz ( 883997 ) on Tuesday September 23, 2008 @06:55PM (#25129395)

    Sadly, while predicting the weather and better understanding it ultimately helps a lot of people, I suspect a LOT more computing power is thrown at more mundane things like predicting where the financial markets are going to be based on a gazillion data inputs. Probably even better funded are the vast datacenters around the world that fondle communications and other data for the spymasters. I doubt those computing resources are represented in the annual supercomputing lists. :)

    • by Quarters ( 18322 ) on Tuesday September 23, 2008 @07:03PM (#25129469)
      But, in a roundabout way the financial market simulator will ultimately help the weather simulator's performance. Everyone knows that business apps are written in VB. That means the financial simulator folks need MUCH more powerful supercomputers to run their code at anything close to appreciable speed. That same machine will run well coded weather apps blazingly fast!
    • Re: (Score:2, Insightful)

      by blair1q ( 305137 )

      I suspect no computing power is being thrown at predicting where the financial markets are going.

      A lot is thrown at pretending to predict it, but it's brilliantly obvious that the output of such things is no better than chicken entrails or the last two presidential elections.

    • 90% of Roadrunner CPU time is reserved for the military as I recall.
    • I suspect a LOT more computing power is thrown at more mundane things like predicting where the financial markets are going to be based on a gazillion data inputs.

      I wouldn't say mundane.

      For one thing, any intelligent deep thought requires the gazillion inputs because of the sheer number of factors and an error or bad approximation in just a few inputs could throw off the decision or computation. Granted, the financial markets are typically so frenzied that hardly anyone knows whether to buy or sell, but we

    • Re: (Score:3, Interesting)

      Sadly, while predicting the weather and better understanding it ultimately helps a lot of people, I suspect a LOT more computing power is thrown at more mundane things like predicting where the financial markets are going to be based on a gazillion data inputs. Probably even better funded are the vast datacenters around the world that fondle communications and other data for the spymasters. I doubt those computing resources are represented in the annual supercomputing lists. :)

      There are a couple of misperceptions here.

      Both the problems described, modeling years of weather models or modeling financial instruments, suffer from a definite flaw: they are not mathematical problems in the "high school" sense of the world, i.e. it is not possible to prove that there is only one finite solution that is demonstrably right.
      Financial models are "fit to reality": you take a long time series, make a few wild guesses, throw it into a Cray-2, and look what the model says. Lather, rinse, r

      • by jstott ( 212041 )

        Weather modelling in yearly time scales suffers from the same flaw, in my view: unless you have a long enough set of possible inputs , it's not verifiable in reality.

        Weather modeling (what the National Weather Service does) and climate modeling (what global warming folks do) are different problems.

        Weather modeling has lots of data available (365 forecasts per year, per model pixel) and its easy to verify predictions against weather station measurements. The problem is that it's impossible to specify the c

  • Flops not useful? (Score:5, Informative)

    by zippthorne ( 748122 ) on Tuesday September 23, 2008 @06:57PM (#25129413) Journal

    But.. The whole point is to test the model, and the models change, don't they? Surely we're not just simulating more "years" of climate with the current batch, but improving resolution, making fewer simplifying assumptions, and hopefully, finding ways to do the exact same operations with fewer cycles.

    How can you possibly evaluate supercomputers in any other way except how many mathematical operations can be performed in some reference time? And.. some serial metric if the math is highly parallel, since just reducing the size of vectors in those cases wouldn't actually result in those flops being useful for other tasks.

    • Re:Flops not useful? (Score:5, Interesting)

      by geekoid ( 135745 ) <{moc.oohay} {ta} {dnaltropnidad}> on Tuesday September 23, 2008 @07:03PM (#25129473) Homepage Journal

      That's just the problem, people want to hear raw numbers, but those are useless.
      How well can it do the specific task it needs to do is the actual question. It's a hard one, to be sure.

      • Re:Flops not useful? (Score:5, Informative)

        by jd ( 1658 ) <imipakNO@SPAMyahoo.com> on Tuesday September 23, 2008 @08:31PM (#25130127) Homepage Journal
        To be honest, I thought most people already knew about and used HPC Challenge, which produces 7 different benchmarks covering different types of mathematical problem, memory bandwidth and communications bandwidth. I also imagined people would use MPI-IO [lanl.gov] for measuring MPI performance, that the numbers on the Top500 was simply because it's hard to track a vast number of stats in a meaningful way.

        Of course, if it's actually the case that people are dumb, lazy or in marketing, then that would explain why we don't get a full range of stats, even though the tools have existed for many years and are certainly widely known.

    • Re:Flops not useful? (Score:5, Informative)

      by corsec67 ( 627446 ) on Tuesday September 23, 2008 @07:13PM (#25129573) Homepage Journal

      Flops wouldn't test how well the interconnects work.

      Since you say "increase the resolution of the model", you are expanding the size of the model, and how much data must be used by all of the nodes of the computer.

      Since how important the interconnect properties are is dependent on the model, with almost no communication needed, like for F@H, to a problem that needs all of the nodes to have access to a single shared set of data, it would be very hard to quantify performance in one number.

      Unfortunately, there are more than a few fields where marketers want a single number to advertise in a "mine is bigger than yours" competition, and come up with a metric that is almost worthless.

      • by jd ( 1658 )
        Oh, and you can perform further diagnostics by using DAKOTA and KOJAK to profile a standard MPI application across the supercomputer. Since these profilers give you internal details, you should be able to see where specifically slow-downs occur - be it a processing quirk, a system library flaw or a networking glitch. Hell, I wish someone paid me to benchmark supercomputers. I can pretty much guarantee RoleMaster fans will love the numbers, even if nobody else can cope with the stats.
      • Re: (Score:3, Informative)

        by ceoyoyo ( 59147 )

        Actually, Linpack is not embarrassingly parallel so it DOES test how well the interconnects work, to some extent.

        The top 500 list is interesting, but if you're building a supercomputer to make a certain rank you have too much money and you should really give me some.

        You build a supercomputer to perform some task or class of tasks. If it gets you on the list, cool.

    • by Fishbulb ( 32296 )

      Because the climate simulation model they use does a LOT of inter-process communication. Each piece of the calculation depends on what's going on around it.
      Ever see footage of manual calculation rooms NASA used to have*? Imagine if every one of the calculations those people were doing depended on the previous calculation they did, AND all of the previous calculations of their eight nearest neighbors.

      Now you know why that atmospheric model has a benchmark rated in "century/months" - the number of centuries

    • Re:Flops not useful? (Score:5, Informative)

      by Salamander ( 33735 ) <jeff.pl@atyp@us> on Tuesday September 23, 2008 @07:46PM (#25129839) Homepage Journal

      How can you possibly evaluate supercomputers in any other way except how many mathematical operations can be performed in some reference time?

      Simple: you evaluate how much actual work it can perform across the entire system per unit time, where "actual work" means a mix of operations similar to some real application of interest. The whole problem here is that practically no real application is as purely focused on arithmetic operations as Linpack. Even the people who developed Linpack know this, which is why they developed the HPCC suite as its successor. It's composed of seven benchmarks, including some (e.g. stream triad) that mostly stress memory and some (e.g. matrix transpose) that mostly stress interconnects. If you want to get an idea how your application will perform on various machines, you determine what mix of those seven numbers best approximates your application, assign appropriate weights, and then apply those weights to the vendor numbers. Then you negotiate with the two or three most promising vendors to run your application for real. SPEC should have put an end to simplistic "single figure of merit" comparisons, or if not them then TPC, SPC, etc. Sadly, though, there's still always someone who comes along and tries to revive the corpse.

      • The hard part of that is that my hypothetical high-end supercomputer application and your hypothetical high-end supercomputer application don't do the same thing, and so we must weight those values differently.

        To get a single unbiased score, there must be one simple test or a group of unweighted ones. If you're giving an equal weight to different benchmarks, then the overall score might still mean something. The single value from equal weighing still won't be very predictive for applications that stress one

      • "Simple: you evaluate how much actual work it can perform across the entire system per unit time, where "actual work" means a mix of operations similar to some real application of interest"

        Personally I'd like to see supercomputing power start automating detecting bad design and potential failures of someone elses code, not to mention optimization.

        I've always wondered what you could do with google and a few supercomputers coming the open source database where you can deduce what a function does without needi

    • Re: (Score:3, Interesting)

      by rockmuelle ( 575982 )

      "How can you possibly evaluate supercomputers in any other way except how many mathematical operations can be performed in some reference time? "

      It's much more subtle than that. Most programs, including weather simulations, use a large amount of data stored on disk and in RAM. The problem with LINPACK as a benchmark is that, for all practical purposes, it ignores this cost by using a few very specific linear algebra operations that have very low communication/computation ratios. The LINPACK number is onl

    • Re: (Score:2, Funny)

      by RulerOf ( 975607 )

      But.. The whole point is to test the model, and the models change, don't they?

      Alas, no. It's suspected that the closer your computer climbs to the top of this list, the larger your penis becomes.

  • by straponego ( 521991 ) on Tuesday September 23, 2008 @07:02PM (#25129459)
    A quality HPC vendor will give you the opportunity to benchmark your application before you buy a system or cluster. Most will have standard codes installed, but you should also be able to arrange for a login to build and run your own code on their test clusters. This is the only way to guarantee you're getting the best bang per buck, because the bottleneck in your particular applications may be memory, IO, interconnect, CPU, chipset, libraries, OS... An HPC cluster can be a big purchase, and it performance and reliability can make or break careers. Don't trust generalized benchmarks unless you know that they accurately reflect your workload on the hardware you'll be purchasing.
  • I agree (Score:3, Informative)

    by friedmud ( 512466 ) on Tuesday September 23, 2008 @07:04PM (#25129479)

    I write massively parallel scientific code that runs on these supercomputers for a living... and this is what I've been preaching all along.

    The thing about RoadRunner and others (such as Red Storm at Sandia) is that they are special pieces of hardware that run highly specialized operating systems. I can say from experience that these are an _enormous_ pain in the ass to code for... and reaching anything near the theoretical computing limit on these machines with real world engineering applications is essentially impossible... not too mention all of the extra time it costs you in just getting your application to compile on the machine and debug it...

    My "day-to-day" supercomputer is a 2048 processor machine made up of generic Intel cores all running a slightly modified version of Suse Linux. This is a great machine for development _and_ for execution. My users have no trouble using my software and the machine... because it's just Linux.

    When looking at a supercomputer I always think in terms of utility... not in terms of Flops. It's for this reason that I think the guys down at the Texas Advanced Computing Center got it right when they built Ranger ( http://www.tacc.utexas.edu/resources/hpcsystems/#constellation [utexas.edu] ). It's about a half a petaflop... but guess what? It runs Linux! And is actually made up of a bunch of Opteron cores... the machine itself is also a huge, awesome looking beast (I've been inside it... the 2 huge Infiniband switches are really something to see). I haven't used it myself (yet), but I have friends working at TACC and everyone really likes the machine a lot. It definitely strikes that chord between ultra-powerful and ultra-useful.

    Friedmud

    • Re:I agree (Score:4, Funny)

      by Abreu ( 173023 ) on Tuesday September 23, 2008 @07:24PM (#25129663)

      My "day-to-day" supercomputer is a 2048 processor machine made up of generic Intel cores all running a slightly modified version of Suse Linux.

      We all envy you.

    • Well, let's see (Score:5, Interesting)

      by Louis Savain ( 65843 ) on Tuesday September 23, 2008 @07:29PM (#25129715) Homepage

      It's about a half a petaflop... but guess what? It runs Linux!

      This sounds kind of nice but why should this make it any easier to write parallel programs for it? You still have to manage hundreds if not thousands of threads, right? This will not magically turn it into a computer for the masses, I guarantee you that. I have said it elswhere [blogspot.com] but parallel computing will not come of age until they do away with multithreading and the traditional CPU core [blogspot.com]. There is a way to build and program parallel computers that does not involve the use of threads or CPUs. This is the only way to solve the parallel programming crisis. Until then, supercomputing will continue to be a curiosity that us mainstream programmers and users can only dream about.

      • Re: (Score:3, Insightful)

        by friedmud ( 512466 )

        "Until then, supercomputing will continue to be a curiosity that us mainstream programmers and users can only dream about."

        I'm not so sure that's a bad thing... most applications don't need the power of a super computer...

        At the same time, I agree that I wish that desktop development tools made it easier to do threading for multi-core machines. Every new computer comes with more than one core... but the development tools (languages, compilers, IDE's, debuggers) simply aren't helping the everyday joe progra

      • First off... I hate threads... they are a locking nightmare once things get complicated,, finding a halting problem isnt worth the effort.

        Really I'm an LAM-MPI freak.. let all the processes talk, gather data and not share memory. it skips the pthreads issue, but some would call the mpirun a thread-launcher.

        Nope, mpi wont make parallel code a computer for the masses... I dont have ANY clue as to how to pull the CPU out of the equation.

        However there are a BUNCH of ways that parallel code can be commo

      • by dkf ( 304284 )

        This sounds kind of nice but why should this make it any easier to write parallel programs for it? You still have to manage hundreds if not thousands of threads, right? This will not magically turn it into a computer for the masses, I guarantee you that. I have said it elswhere [blogspot.com] but parallel computing will not come of age until they do away with multithreading and the traditional CPU core [blogspot.com]. There is a way to build and program parallel computers that does not involve the use of threads or CPUs. This is the only way to solve the parallel programming crisis. Until then, supercomputing will continue to be a curiosity that us mainstream programmers and users can only dream about.

        You're wildly mistaken, but for interesting reasons. While SIMD architectures are indeed interesting and do very well on some types of problem (I believe that weather simulation is good) on other problems it is far easier to express them as either MIMD or something more exotic. A real system that needs to handle a mix of codes needs to be implemented as a hybrid: e.g. a cycle-scavenged Condor pool for high-availability, a traditional cluster with a better interconnect, and a supercomputer for those times wh

      • by jstott ( 212041 )

        You still have to manage hundreds if not thousands of threads, right?

        Don't think of these things as 2048-core CPU's sharing memory and disk space, think of them as clusters of 2048 separate [asnychronous!] PC's, each with its own memory and disk space. The actual architectures will vary, but this is usually the right paradigm (Crays and other big vector processors are the main exception to this rule). Think, then, of the program as if it were a bunch of network clients talking to each other over sockets

    • RoadRunner (Score:2, Insightful)

      by Anonymous Coward

      Is running relatively stock Fedora (the ppc distribution). True, it's ram hosted, but the OS is hardly specialized in terms of libraries and such. You could say the Cell SDK is a tad specialized, but the underlying platform is not so custom as implied.

      In fact, every single Top500 system I've ever touched has been far more typical linux than most people ever expect.

      In any event, the most compelling aspect of RoadRunner in my view is the flops/watt. Application developers who can leverage highly parallel c

      • Re: (Score:3, Informative)

        by friedmud ( 512466 )

        The specialization of the hardware / software combo is what I was referring to.

        Have you ever coded for one of these special architectures? It really is a bitch. Yes, Redstorm is even worse (special OS that doesn't even allow dynamic linking!)... but the non-generality of the cell-processors is going to kill the real world impact of Roadrunner.

        ASCII Purple was one of the previous machines at LANL that was a "one-off" build from IBM. It was a complete disaster. Porting code to the machine took much longer

        • Re: (Score:3, Informative)

          by friedmud ( 512466 )

          Sorry... got my supercomputers mixed up... ASCII Purple was at LANL...

          I was thinking of ASCI Q, but that was made by HP...

          Oh... just nevermind... I screwed it up well enough, just forget it ;-)

          Need to get some sleep.

          Friedmud

        • by joib ( 70841 )

          Hasn't Red Storm switched to Linux yet? Seems most of the Cray XT sites have done so already. One reason for developing CNL (Compute Node Linux, the Linux kernel they run on the nodes) IIRC was that multicore support in Catamount was sort of a kludge. And OS buffering, and..

          Then again, the catamount developers are at Sandia, so maybe they have some emotional attachment to it. :)

          CNL still doesn't allow dynamic linking, though it's much less alien than catamount.

          • Re: (Score:3, Informative)

            by friedmud ( 512466 )

            As far as I know (as of 3 months ago) they're still running Catamount at Sandia... and it's for the reason you state: they developed it.

            Friedmud

    • Re:I agree (Score:4, Insightful)

      by CronoCloud ( 590650 ) <[cronocloudauron] [at] [gmail.com]> on Tuesday September 23, 2008 @07:36PM (#25129765)

      The thing about RoadRunner and others (such as Red Storm at Sandia) is that they are special pieces of hardware that run highly specialized operating systems.

      from http://en.wikipedia.org/wiki/IBM_Roadrunner [wikipedia.org]

      The Roadrunner uses Red Hat Enterprise Linux along with Fedora as its operating systems and it's managed with xCAT distributed computing software.

    • Not just buying the thing, also the cost of maintaining it (standard hardware is most likely easier to maintain), and the power it and its cooling system uses, check green500 for that one. Actually, as a user, I have often found that most supercomputing clusters are inefficient for at least the first year-and-a-half due to imperfect queuing systems or network/filesystem incompatibilities. "Yeah, your run will likely crash every now and then but we don't know yet why". I do not blame the administrators, I b
    • by blair1q ( 305137 )

      Your users don't really need that kind of power, then. They could get by using much less of a computer, because they're spending less time coding and have time for running. In fact, I suspect about 70% of your users could get by on a 4x4 card if they ran it and went on vacation for the summer.

      The really big HPCs can tolerate the really big software development efforts because the runtime saved by the specialized OS may more than discount the extra development effort.

      • "he really big HPCs can tolerate the really big software development efforts because the runtime saved by the specialized OS may more than discount the extra development effort."

        The thing is... that might be true for a very select number of projects... and that number of projects is actually smaller than most people think.

        Yes, my code can run up to 3 or 4 thousand processors nicely... to get some really high fidelity engineering simulations... _but_ most people are running in the 64 to 128 processor range d

  • Non-story... (Score:3, Insightful)

    by Kjella ( 173770 ) on Tuesday September 23, 2008 @07:05PM (#25129497) Homepage

    ...ever looked at gaming benchmarks? Server benchmarks? Productivity benchmarks? Rendering benchmarks? In fact, any kind of benchmark? Seen how they all differ depending on the product and test run? Same with supercomputers, you got some synthetic benchmarks, and you got some real world benchmarks. But the weather simulation may not be a relevant benchmark at all if you're doing nuke simulations or gene decoding or finite deformation or some other kind of simulation. Synthetics are the lowest common denominator - you'd rather see benchmarks in your field, and most of all benchmarks with your exact application. That doesn't change that those are individual wants and synthetic benchmarks are the only ones with any value to everyone.

    • "...ever looked at gaming benchmarks? Server benchmarks? Productivity benchmarks? Rendering benchmarks? In fact, any kind of benchmark? Seen how they all differ depending on the product and test run? Same with supercomputers, you got some synthetic benchmarks, and you got some real world benchmarks."

      I hear what you're saying, but if we look at application complexity (modern games for instance) there really there isn't any such thing as "real world benchmarks" since each application is specific because of t

  • by Junta ( 36770 ) on Tuesday September 23, 2008 @07:23PM (#25129657)

    Just with a lot more dollars behind it...

    Every one remotely engaged in Top500 systems knows how very specific the thing being measured is. It's most sensitive to the aggregate clock cycles and processor architecture, and not as sensitive to memory throughput/architecture or networking as many real world things are.

    http://icl.cs.utk.edu/hpcc/ [utk.edu]

    Is an attempt to be more comprehensive, at least, by specifying a whole suite of independently scored benchmarks to reflect the strengths and weaknesses of things in a more holistic way. Sure, it's still synthetic, but it can give a better 'at-a-glance' indicator of several generally important aspects of a supercomputer configuration.

    The thing probably inhibiting acceptance of this is that very fact, that it is holistic and the winner 'depends' on how you sort the data. This is excellent for those wanting to more comprehensively understand their configurations standing in the scheme of things, but hard for vendors and facilities to use for marketing leverage. Being able to say 'we built *the* fastest supercomputer according to the list' is a lot stronger than 'depending on how you count, we could be considered number one. Vendors will aggressively pursue pricing knowing about the attached bragging rights, and facilities that receive research grant money similarly want the ability to make statements without disclaimers.

    Rest assured, though, that more thorough evaluations are done and not every decision in the Top500 is just about that benchmark. For example, AMD platforms are doing more strongly than they would if only HPL score is counted. AMD's memory performance is still outrageously better than Intel and is good for many HPC applications, but Intel's current generation trounces AMD in HPL score. Of course, Intel did overwhelmingly become popular upon release of their 64-bit core architecture based systems, but still..

  • by Anonymous Coward

    Most of the locations listed are mostly educational institutions, r&d centers, and computer companies. The results were probably submitted unofficially. There are few exceptions, but they are just that--few. It makes you wonder what the Big Data companies (Google, Yahoo!, etc) actually have running. They have no reason to participate, after all...

    Consider something like Yahoo!'s research cluster [yahoo.com]. Why isn't it on this list? Why don't they run the tests?

  • That an article about featuring IBM supercomputers comes shortly after a few misguided individuals were posting that "IBM is no longer relevant, they are a OEM reseller nowadays" or that they "only make bloated, slow software"

  • Number of simulated years per day isn't exactly the metric you want. I can simulate a million years in a minute on my home pc, just not very accurately. As you get more accurate, the sim years/CPU day will decrease.

    So knowing the number of simulated years per cpu day doesn't tell you anything unless you know exactly what algorithm you're using.

  • by Raul654 ( 453029 ) on Tuesday September 23, 2008 @08:56PM (#25130269) Homepage

    It's fair to criticize Linpack for being a one-trick pony. It measures system performance for dense linear algebra, and nothing else. Jack Dongarra (the guy who wrote Linpack and maintains the top 500 lists) is quite up-front about Linpack's limitations, and he thinks that using a single number as the end-all-be-all of a computer's performance is a bad idea. It's a simple fact of life that certian kinds of computers do better on certain problems. The good guys out at Berkeley even sat down a couple years ago and enumerated [berkeley.edu] all of the problems they found in real-world HPC applications (See the tables on pages 11-12). The real truth here is that people should stop treating Linpack like it's the final measure of system performance. If you are doing pure linear algebra problems, it's a pretty good measurement for your purposes; if you are not, then you use it at your own peril.

  • They should just run Vista on them to use as a benchmark. That will effectively flood bottlenecks on all kinds of levels.
    • What's even funnier is the reality that Windows can't even boot on that many CPUs, let alone scale to utilise them.

  • The thing about climate models is that they get more complex and higher resolution as soon as the computers get faster. We will always take about 3 months to run a simulation. You can run it faster? Make it more detailed. It takes longer? Wait a few months to a year and it'll only take 3 months. Why 3 months? Not sure. Partly because that is about the length of a funding cycle of design experiment, run it, analyze, and write it up.

    If you want to run 100 years per day, you can do so with an older model. T
  • I seem to recall a Nova special I watched many moons ago about "strange attractors" and "fractal behavior" that seemed to indicate that for a large class of complex-valued iterative functions there was a weird phenomenon called the "Butterfly Effect". Apparently... according to this show I saw 20 years ago (and I think that Mandelbrot mentioned it in a lecture I attended a few years later), initial variables which are as intertwined as the rational and irrational numbers can have drastically divergent outco

    • by azgard ( 461476 )

      Well, it's actually quite possible to model climate but not weather. It's an issue of scales. I won't go into mathematical argument, because I don't understand it anymore, and you probably don't want to know that either.

      But look at this thing: http://en.wikipedia.org/wiki/Lorentz_attractor [wikipedia.org] - it is essentially an infinite line curved into a strange "two ears" shape. The line is defined by 3 ordinary differential equations, and it is completely defined by selected point and direction in space.

      The interesting

  • Synthetic benchmarks aren't applicable to all users! Who knew?
    TZ
  • I'll save those huge computers a few million years: The answer is 42.

    Now try producing that on a mere Milliard Gargantubrain.

  • Software is the weak point in supercomputing nowadays. Funding for hardware SO outstrips the development of software that runs on the expensive hardware that it's becoming a SERIOUS problem.

  • State-of-the-art systems today can simulate about five years per day of computer time, he says, but some climatologists yearn to simulate 100 years in a day."

    Something just doesn't jib here. So it takes you 20 days instead of 1 to compute the next hundred years. Why is this a problem? It's still a hundred years in 20 days. Don't you have even a little bit of patience?

    The only reason to compute a hundred years in one day is if you're going to restart the computation each morning to see what the next

  • Moving from PDPs to VAXen. Doesnt seem all that long ago.
    About 13, 14 years for each new 1000x level.
  • "State-of-the-art systems today can simulate about five years per day of computer time, he says, but some climatologists yearn to simulate 100 years in a day."

    Strangely enough, I'm reminded of the Heisenberg uncertainty principle. Considering the power and infrastructure requirements to run a supercomputer with that much processing power today, you might change the climate by simulating it.

Some people manage by the book, even though they don't know who wrote the book or even what book.

Working...