Swarm — a New Approach To Distributed Computation 80
An anonymous reader writes "Ian Clarke, creator of Freenet, has been working on a new open source project called Swarm. The concept is to allow a computer program to be distributed across multiple computers in a manner almost completely transparent to the programmer. The system observes the program executing and figures out how the workload should be distributed for maximum efficiency. Swarm is implemented in Scala. Its at an early-prototype stage, and Ian has created a good 36 minute video explaining the concept and the current implementation."
Re: (Score:3, Funny)
Re:This'll be great for botnets (Score:5, Insightful)
Re: (Score:2, Interesting)
Re: (Score:2)
Re: (Score:2, Informative)
My Dell desktop from 1999 has been running like the wind again since last week, when I reverted it to its 2002 state from backup tape. It goes superfast now that it's virus-free, off the network, and running old apps on Windows 98.
I was only trying to recover some old files before junking an unusable m
Re: (Score:2)
Re: (Score:3, Interesting)
That's true to some degree. But computers do slow down as they age. Components damaged by the constant heating cause more errors and therefore require retransmission or error correction, slowing things down.
No, not really. PCs are nowhere near that sophisticated. A high-speed CPU bus is not like a DSL connection. Pretty much it has to work near-perfectly, or it's blue-screen city.
... still as fast as it ever was (faster, actually ... I have it running a stripped-down version of XP.) If you have a motherboard or PC that is getting erro
For example, I have a couple of Athlon 1.4 ghz machines that are running just as fast as the day I built them, and they've never been turned off. Also have an old Thinkpad R41
Re: (Score:2)
So long story short it is time that we get some proper software running on those computers.
For 99% of the people a computer is an appliance, like the TV and the stereo. They do not get more maintenance other than being dusted off once in a while.
Defragging harddrives: is that still necessary in the Windows world? I stopped doing this more than 15 years ago, at the time running OS/2 and its HPFS.
Getting infected: yes that's an issue and I have honestly no idea how to really prevent this. Even a fully lock
Re: (Score:2)
People's machines slow down because a. they never defrag their hard drives and b. they get infected.
You also need to take into account that they may install new service packs and other software to tread water, not to mention adding new bloatware. For example, probably hundreds of thousands of PC desktops and laptops were sold with Windows XP Home and 256MB of RAM, that were not slow at the time. But try running them today with Service Pack 3 and adding antivirus (like AVG) and firewall (like ZoneAlarm) programs, without adding any RAM, and they're terrible. Now consider that many of those same systems als
Re: (Score:2)
That's not really bad. It moves the PC market forward.
Re: (Score:2)
Learn the broken window fallacy, [wikipedia.org] please.
Re:This'll be great for botnets (Score:4, Insightful)
In my experience, Java is not the reason people buy new computers.
Their computers slow down from viruses, or virus-like Antivirus, and then they think they need to upgrade.
Lately commercially made programs (AIM? Windows Live stuff? Most printer software? Most shareware?) seem to consume as much memory as a whole JVM, despite being written in C. This has led me to conclude that companies really don't give a shit how much memory their software uses. This is quite ironically pushing Java closer and closer to C in actual memory and CPU usage.
Disclaimer: I know C is amazing when used properly - but it seems like only small FOSS projects and apps destined for phones have any sort of optimization work done. I've seen daemons use 200KB on a tiny linux handheld, but multiple megabytes is the norm on any desktop.
Re:I'd pick ont the $NtServicePacks (Score:1)
Re: (Score:3, Funny)
Nope, I already use OpenOffice!
Re: (Score:2)
I'm sure you would notice an apparently suspicious huge JVM process eating your CPU time. :]
How is that different from any other kind of JVM process?
Earlier (Score:5, Interesting)
.. was Mosix http://www.mosix.org/ [mosix.org]
It allowed mosix-running linux computers to distribute their loads over a connected other mosix-running linux computers.
Processes migrate to other nodes transparently. No programming changes were needed.
Re:Earlier (Score:4, Informative)
Re: (Score:2)
One key component of Swarm is that a supervisor process uses a clustering algorithm to determine how data should be distributed such that it minimizes the number of times a continuation must jump between different computers. Does Mosix have any equivalent?
Why has Mosix not achieved wider usage, for example, allowing web applications to scale up using multiple servers?
Re:Earlier (Score:4, Informative)
It's not free software so you can't use it except for personal or educational use. Open Mosix died
http://openmosix.sourceforge.net/ [sourceforge.net]
Re: (Score:2)
Is it wrong to hope that MOSIX dies, and possibly Frees their code? I want a single system image, where when I bring up a laptop all my computers get faster, but I don't want to be stuck with one kernel and 8 nodes (it's for home use, so I could use it but I'd have to track their kernel.)
Another Earlier - ERLANG! (Score:5, Informative)
Erlang apparently gets it right. Scales smoothly from single core to multi-core to multi-server in a near linear fashion. Astonishingly reliable, having achieved nine nines of uptime - much less than a second of downtime - in a year. Purposely designed to mitigate shared memory problems. Built for hot-switchover - you can upgrade Erlang problems without closing them first!
In just about every conceivable way, Erlang is the right choice for high-end multi-core multi-system clustered application development. I have a large-stack, clustered application written in PHP. While it works well, there are limits to what we can do within a single process - a problem that's likely to become worse over time as needs continue to scale up. If I were to do it all over again, I'd take a good, hard, look at Erlang.
Re: (Score:2)
Maybe it's time for you to start looking into Erlang (and alternatives) NOW. Not when your demands get so high your current application breaks down. I suspect this is really complex stuff, very hard to get it right lest to understand what it is doing really in the first place.
Re: (Score:2)
Correct! I have written distributed systems half of my life (longer than the age of maybe most readers here?), relying mainly on 'C', TAL, even Pascal and assembler because of the company requirements but (just for fun) tried it in Erlang - amazing for so old language!
It has about everything you can think and all that in language! Multiple platforms, own transaction / memory/ whatever databases, can (I tried that!) be used with all main languages, easy syntax, small programs, failsafe, etc, etc.
Used by (hug
Re: (Score:2)
- Erlang didn't get less than a second of downtime in a year, an application written in Erlang got less than a second of downtime in a year. I bet people clever enough to write such an application in Erlang could have written it in another language. Would it have been more difficult? Probably. But just because you use Erlang doesn't mean that your application is going to magically never going to have downtime - you're still going to have to work hard at it.
- Erlang is not necessarily the right choice fo
Re: (Score:1)
The good news is that Mosix is still alive and well.
The bad news is that although Mosix is excellent for High Performance Computing (HPC), it is totally useless for data-mining and web applications, which is just what Swarm is all about!
Name... Neat idea though (Score:3, Interesting)
Re: (Score:2)
Re: (Score:3, Insightful)
I'm just getting into Agent Based Modeling myself and I had exactly the same thought... why would they use the name of an established tool; especially when there are similarities in the concepts. This seems like a recipe for confusion.
A good first check when starting an open project is to check propesedprojectname.org and see if there's anything active there. Or even just Google it - if another project shows up near the top with the same name, it's probably a good idea to pick another name.
I'm sure there
Re: (Score:2, Informative)
From the FAQ [google.com]:
Re: (Score:1)
Re: (Score:2)
How about ...
Hurd?
Obligatory (Score:4, Funny)
Re: (Score:1)
Re: (Score:2)
looks intriguing (Score:5, Insightful)
The thing that's always killed this idea (along with automatic parallelization even on the same machine) is that the overhead of figuring out what's worth distributing, and the additional overhead from mistakes (accidentally distribute trivial computations), often swamps the gains from the multiple processors banging away on it simultaneously. Determining statically what's worth distributing is very hard, since solving it properly is undecidable (basically equivalent to the halting problem), and even solving it in a significant enough subset of cases to be useful has proved difficult. It looks like this project is monitoring dynamically to determine what to distribute, which seems likely to be more fruitful, although historically that approach has suffered from the overhead of the monitoring (like always running your code with debugging instrumentation turned on).
I certainly hope he has a breakthrough vs. past approaches, or it could just be that advances in a lot of areas of technology have given him a better substrate on which to build things that naturally mitigates lots of the problems these things used to have (automatic parallelization research started probably ahead of its time, back in the 1970s, so that most academic stuff was killed off by the 1990s after no really knock-down results emerged). It's not entirely clear to me what the killer advance is, though. The particular variety of portable continuations? A good way of easily monitoring computations? Something that makes the data-dependency analysis particularly easy?
Re: (Score:3, Interesting)
That kind of thinking is so 90's. Brute force data mining, as an example means harvest it all and let target groups sort out what they want. It is a waste of time to 'decide'. That's like stopping to inspect every shovel full of ore as it comes out of the ground. All or nothing has been the default for some time now, and this is just a
Re: (Score:2)
Re: (Score:3, Insightful)
Depending on how many cores you have access to, distributing trivial computations may not matter. If we ever start seeing 32 core desktop machines, for example, you start to get to the point where forking could create a realtime speedup even though in absolute terms you've wasted 5 times as many cycles.
Re: (Score:2)
It definitely lowers the bar at how good you have to be, but I'm not sure it makes it irrelevant. Just the overhead of putting computations into some sort of container (thunks of some sort) and getting them back out can get absurd if the computations turn out to be, say, smaller than 100 instructions.
Re: (Score:2)
The appealing thing about this is the problem stems from the way we program. It is really difficult to escape the single-core mindset with the languages we have today. I don't necessarily want to invoke the Wharf hypothesis here, but to some degree our expressive power is limited by the language we are using. Imperative languages that I know lend themselves to single-processor execution. I wonder if new languages specifically designed for multi-core programming would be able to avoid some of the problems yo
Re:looks intriguing (Score:5, Interesting)
A friend of mine did a system like this about ten years ago --- hi, Iain! --- called Flit. It had a number of the same features, although using a custom language; it had some rather interesting concepts, such as asynchronous function calls that would return immediately, spawning a new thread, but return a future: a value whose value was not known yet. Accessing the value would cause the thread to be waited upon.
Unfortunately the killer problem that sunk Flit was that of distributed garbage collection. Collecting data over multiple machines is really, really hard, and he never found a usable approach to make it work. I was very disappointed to see that Swarm's garbage collection is still on the to-do list --- he doesn't appear to have started to think about it yet.
I hope he can make Swarm work --- it's something that we could all definitely use. But there are fundamental theoretical problems that have to be solved first...
Re: (Score:1)
It's not always true that monitoring is a cost: JIT (Just In Time compilation) monitors execution, and has yielded significant speed-ups in Java.
I conjecture that the key to distributed computing will turn out to be wasting resources (inefficiency) in some way that serves the overall goal.
Re: (Score:2)
Sounds good (Score:2, Interesting)
It sounds like a good idea, but I don't think the project is far enough along in this video to warrant a posting. Maybe he was using too much of a trivial example to be appreciated in the video, but his explicitly offloading the task to another computer doesn't appear to be very far beyond standard client server models. If it were already automatically transporting processing between different nodes, it'd be much cooler, but that is not a trivial problem to solve. Deciding what should and what shouldn't
We have just witnessed the birth of a new meme... (Score:4, Funny)
In Ian Clarke's Swarm, World "Hellos" you!
I doesn't do much yet (Score:4, Informative)
Re:I doesn't do much yet (Score:4, Insightful)
Mod parent up. This is exactly what Ian did with Freenet.
He cobbled together an overly-simplistic prototype to address a set of very difficult unsolved problems in anonymous communication and then farmed out the actual real-world legwork on those problems to interested open source developers while Ian himself effectively abandoned Freenet for other (paying) gigs. To this day he is credited, somewhat ironically, as "the creator of Freenet," and a decade later the Freenet project still hasn't solved the problems it set out to solve, even after changing the fundamental network architecture several times.
Great career strategy though. Get credit for the shiny things and pass the shame of failure off on others. He's CEO material all the way.
Re: (Score:3, Interesting)
I don't think this characterization is fair, and I think you would have a hard time finding somebody who actually worked on Freenet to agree with you. Ian's orginal technical ideas for Freenet - as well as his vision - are very much still a big part of the architecture, and he could never be said to have abandoned it. In fact, time has vindicated many of his ideas to a far greater extent than I expected when we started working with them. You are right that the project has not yet solved the problems it set
Re: (Score:2)
I agree. That said, I actually sat through the entire video, and sadly, it was a disappointment. I ignore everything that he did that came before. He has great intelligence, a wonderful idea, some vague but intriguing suggestions on how to accomplish it, and very little working code.
At least he is up front about this - if by up front, you count a litany of near-impossible problems you do not even have an idea about how to solve in the last 8 minutes of your 35 minute video. :) I do hope for the death of the
Re: (Score:2)
I do hope for the death of the fad of using videos where a perfectly good blog post will do.
I thought that's what slashdot comments were for?
Re: (Score:1, Insightful)
Clarke never said he doesn't know how to solve the remaining problems (of which, he freely admits in the video, there are many). Would you prefer that no open source project was released to the world until it was 100% finished? Good luck with tha
Re: (Score:1)
DAMPVM? (Score:1)
36 minutes? (Score:2)
Re: (Score:2)
Alpha code... (Score:4, Funny)
Computer 1: MOV AL...what? No more? MOV AL what? I need a value! WTF am I supposed to do with that!?
Computer 2: 09? Nine? Who gave me nine on its own. That doesn't make any sense! Jeez! Hey, anyone out there missing some data?
Computer 3: Not me, I'm pushing the registers onto the stack
Computer 4: Nope, I've got an INT
Computer 5: Oh, hey, it could be me - does NOP have a value. No? Sorry, my bad!
Computer 1: Nine - yeah, nine - Well, I could stick that in AL if no-one else wants it!?
Computer 3: Oh, heck, give it to 1. I've just got a POP instruction so I am going to obliterate it anyway.....
Yet another language. (Score:2)
I do respect Ian, but cant we do this with the existing language infrastructure and just extend it?
Re: (Score:1)
with the existing language infrastructure and just extend it?
That's exactly what Scala does. It "extends" Java in a sense and still runs on the JVM. Swarm is essentially a creative new way of using a mature platform, assuming it works at some point.
VMware doing it (Score:2)
Isn't that what the new vSphere or some up-and-coming release from VMware supposed to do?
-m
Fix Freenet First (Score:2)
I think he should fix the monstrosity that is Freenet before he jumps onto other things.
Distributed what now? (Score:2)
I read that as "distributed copulation" for some reason. I need more sleep.
Big claims, belied by reality? (Score:1)
"Ian Clark of the Freenet fame".. Actually, practically no claim about Freenet came true. The authors advertised "anonymity" etc. etc. at the same time as university professors published studies of statistics about the snooped connections: to any node present on the network for some time it is elementary to collect IPs.
It was painful to see so many users completely duped by the untrue claims, which their authors knew pretty well were untrue (and of which fact one-word admissions can be found buried somewher
GPU's (Score:2)
Is there a potential to use this on a GPU? The current problem with GPU programming seems to be solved with swarm.
SWARM is good (Score:2)
whatever happened to aglets? (Score:2)
Hmm. Sounds like Amoeba. (Score:2)