Fastest-Ever Windows HPC Cluster 216
An anonymous reader links to an eWeek story which says that Microsoft's "fastest-yet homegrown supercomputer, running the U.S. company's new Windows HPC Server 2008, debuted in the top 25 of the world's top 500 fastest supercomputers, as tested and operated by the National Center for Supercomputing Applications. ... Most of the cores were made up of Intel Xeon quad-core chips. Storage for the system was about 6 terabytes," and asks "I wonder how the uptime compares? When machines scale to this size, they tend to quirk out in weird ways."
Re:finally (Score:4, Interesting)
You've no idea how right you are.
I got to test Server 2008 before it was released to the public. All our internal applications identified 2008 as "Vista".
Clustered Windows Boxes! (Score:5, Interesting)
Only six teras ? (Score:3, Interesting)
So.... six terabytes... isn't that horribly small by today's standards ? I mean, our small backup server here is 2 teras, it's just a cheap PC with a bunch of SATA drives in it.
Does that mean my gaming rig and media server, when combined, constitute an "HPC Cluster" worthy of the top 100 ?
Ghey.
Re:But why?! (Score:2, Interesting)
Okay... (Score:3, Interesting)
But the statistics for the top500.org show that over 9000 processors is way above normal for a supercomputer cluster up there. In fact less than 5% of machines in the entire 500 have more than 8000 processors, with the majority around the 1-4k mark. Oh, and 85% run Linux-only with an amazing 5 (not percent, actual projects) running Microsoft-only. So it looks like MS did this through hardware brute-force, not some amazing feat of programming. But then, that's true of them all. Although being in the top500 list is "good PR", it doesn't mean that much.
I wonder what the licensing is like for a 9000-processor Windows Server, though?
What is the benefit of Windows on a cluster? (Score:3, Interesting)
Can someone explain why anyone could possibly want Windows on a scientific computing cluster? What does Windows offer that Linux doesn't?
Much of my work involves running molecular dynamics simulations. By HPC standards these are tiny calculations (in my case, usually 32 CPUs at a time). All science HPC software I'm aware of is Unix-oriented, and everything runs on Linux. At my institution we have an OS X cluster and we are in the process of purchasing a Linux cluster. We didn't even consider Windows - given the difficulties we've experienced administering Windows on the desktop, a Windows cluster just seems like an expensive exercise in frustration.
Re:What is the benefit of Windows on a cluster? (Score:3, Interesting)
Re:But why?! (Score:3, Interesting)
It's growing yes but its actually a very low margin market. The whole idea of an HPC cluster is saving money.
Somehow I doubt it's the margins so much as the fact that Linux dominates it and they are afraid Linux will use that to gain a foothold elsewhere.
Re:*yawns* (Score:2, Interesting)
Before everyone completely dismisses this story... (Score:5, Interesting)
While I don't agree that Microsoft Windows HPC Server is the best software to manage a supercomputer, the linux diehards out there should pay attention to a problem that Microsoft is trying to tackle: accessible supercomputing. See one of their case studies [microsoft.com] as an example.
The bottom line is, these days pretty much anyone has access to a few TFlops of compute power, but the learning curve for getting something running on these machines is pretty intimidating, especially for non-CS based disciplines. I've had to take a 1-2 day class, plus futz around with the clunky command-line tools for a few days or so, on every supercomputer I've used, just to get simple jobs running. In my experience, people learn to game the various batching and queuing systems such that their jobs run faster than everyone else's, further shutting out the newcomers.
HPC vendors would be wise to focus more attention on the tools and interfaces so that Joe-researcher can set the number of nodes and go, rather than having to manually edit loadleveler text files, sending them to the queue, and then coming back next day to find the job failed due to a typo in the startup script.
On multi-TFLOP systems, not everyone needs 99.5% efficiency with all the implementation details that requires. These days, many people just want their job to run reasonably quickly, with no fuss.
The same thing happened several years ago with the move to high level languages like Python and Ruby. Sure, they're slower than C++ and FORTRAN. But for the vast majority of applications, you wouldn't know the difference on modern processors. And the turn around time and user-friendliness on these languages is so much better, using them is a no-brainer.
Hopefully Microsoft can spur the industry in this direction.
humph..... (Score:3, Interesting)
Re:Before everyone completely dismisses this story (Score:4, Interesting)
From your case study:
"""
In addition, it is investigating ways to allow users to connect remotely to the cluster. It expects to complete the project and move the cluster into production by March 2009.
"""
By time the cluster in the case study allows users to remotely log in, the hardware will have lost at least 1/2 of its value.
While more work is needed to make things user friendly, you have to remember that the funding is there for CPUs; not many folks are forward looking enough to realize researchers really need funding into making stuff easier.
more similar (Score:5, Interesting)
Whereas Server 2008 and Vista share a tad more of their code base.
and *that* is relevant.
And could be humorously be alluded to because of the mis-detection of some software.
accessible supercomputing .. (Score:4, Interesting)
Assuming MS was responding to this imagioned problem
"The contest showed that supercomputers
"but the learning curve for getting something running on these machines is pretty intimidating, especially for non-CS based disciplines. I've had to take a 1-2 day class, plus futz around"
You actually programed a supercomouter - cool. What type and where exactly? How does HPC Server differ in respect to other solutions?
"the Blue Gene family of supercomputers has been designed to deliver ultrascale performance within a standard programming environment [ibm.com]"
"Hopefully Microsoft can spur the industry in this direction"
You mean like continually inventing Apple, badly
One big difference... (Score:3, Interesting)
However, MPI itself has serious issues. The master copy of the program starts/stops slave programs vis SSH, although some implementations also support inetd-style starts. Messages sent to multiple machines are sent sequentially, rather than via a reliable multicast, wasting bandwidth and wasting CPU cycles.
Re:finally (Score:3, Interesting)
What was removed ?
Honestly, most of what bothers me are UI changes that didn't need to be made and in any case make the UI worse, not better.
That said, the only feature removed that comes to mind immediately is the File Types association dialog box from the Folder Options control panel / dialog. In every version of Windows you've been able to add/change file verbs and actions as well as do things like change the icon, description, etc. This gave you a very fine level of control and it was great for those who wanted/needed to use it. In Vista that dialog/tab was nuked and replaced with some vanilla "what program do you want to open this file with" crap.
There are more that I can't recall offhand, but that's probably the biggest. Personally, I think it almost comes across as an insult to Windows administrators that they'd just go off and remove something like that.