Will PPC Become the Preferred Linux Platform? 276
grunkhead writes "Stephan Somogyi, aka the Darwinist, at MacWeek has an interesting take on Linux on the PowerPC in the wake of IBM's release of a free motherboard design for the G3/750, suggesting the PPC could be the preferred Linux platform of the future. "
Re:RISCy statement (Score:1)
RISC : An advertising term concocted by a group of researchers at Stanford in an attempt to positively differentiate their new processor design ideas from the complex, powerful designs popular in commercial machinery.
CISC : An advertising term concocted by a group of researchers at Stanford in an attempt to negatively differentiate popular, more powerful commercial processor designs from those of their new, small instruction set designs.
Comment: Eventually RISC processors exceeded the power of their so-called "CISC" counterparts, however it took time and significant research $$$ to make up for the headstarts that the then popular "CISC" ISAs enjoyed. Both terms now enjoy any real meaning only in advertising departments and processor design cults.
Not really (Score:2)
Bang for the buck.. (Score:1)
How does it compare to say SPARC or something?
Re:RISCy statement (Score:1)
Re:USB support (Score:1)
I just bought an ADS USB card and a Mouseman Wheel to replace my aging Mouseman (original version, USB). Works great on my 9500. Try one; you'll like it.
(As always, #include <std-disclaimer.h>. Moof.)
Re:USB support (Score:1)
Way to be! (Score:1)
-A.P.
--
"One World, One Web, One Program" - Microsoft Promotional Ad
Re:And the answer is: Nope. (Score:1)
The RHWM's reasoning is flawed. They calculate the percentage of RHAT's market value created by "the community" by the percentage of the Red Hat distribution code created by that "community." This is obviously a ludicrous measurement, as the majority of the value of Red Hat is due to their management and marketing, not to their code. If it was just the code, everybody would be buying Slackware and Debian GNU/Linux, and Red Hat would have no value.
Re:not the only one, but.. (Score:2)
Re:Only if Linus comes down to earth again. (Score:2)
A program that won't compile without the end-user manually editing its source-code is a broken program.
It's something the mac comunity cares about. (Score:1)
It's not like I'm going to program for the OS or anything; I use it. The MacOS is by far the most *usefull* and *usable* OS out there. I mean what can I say, you just gotta try it.
As for the hardware. It's good, very good, but that's not the reasone I use a Mac. In fact this good hardware is a bit of a pain since it's typically requires a greater capital investement.
A PPC base linux box is not in the cards for me because Linux is generaly not very usefull. My guess is a G3 based PPC box would be much of a threat to Apple's Mac sales since it couldn't run the MacOS. (as it currently stands).
Re:why would macworld publish this? (Score:1)
No way in hell.
Even if you did have Apple ROMs and ASICs you're still doomed since the MacOS isn't hardware abstracted so any small changes in hardware require a new version of the MacOS.
Why do you think the computers that are introduced after a the latest version of the MacOS have enablers? -> the reasone? Apple hasen't included support for that model in the MacOS yet.
fun, eh?
Linux will keep processor makers on their toes. (Score:2)
--
Re:Only if Linus comes down to earth again. (Score:1)
RISC vs CISC is meaningless (Score:1)
The "breakthrough" of the RISC design wasn't coming up with simplier instructions, it was breaking the CPU into independant subsystems that could work in parallel. Doing an integer operation at the same time as doing a floating point one, while loading data from the bus/memory means 3 things are going on at once. A "simple" instruction set means that it is more obvious where the pieces are, but is fundamentally irrelevant.
When programming in assembly (the only time RISC/CISC is visible to a programmer), you want it to be as CSICy as possible, it makes your life easier. Think about
C = A + B
In an old assembler that is one instruction. Simple, easy to read (for assembly). In a load-and-store system that is
load A to r1
load B to r2
add r1 to r2
store r2 to C
4 vs 1. You tell me which one you want to hand code in.
Also, RISC pushes more of the work to the software, which is fundamentally slower than hardware. Lets do as much stuff as low as possible so our systems run faster. Adding layers sucks.
Imagine a future with CISC instructions with pipelined cores. How one gets translated to the other is meaningless to programmers although an interesting research topic for the hardware folks. Think of it as a library. You just care what the interface is (the opcodes), let the library designer (EEs) handle the details of getting that interface to work well. Maybe some of the ease of the monster CISC stuff needs to go away to help out the core (trade-offs are part of an engineer's life), but as a rule CISC is better for programmers.
Just lest RISC vs CISC be forgotten. We need bits and pieces of each one.
Re:this arg is meaningless (Score:1)
You are absolutely right that if a compiler can break down an instruction into smaller pieces, the hardware has less work to do. The is pushing some of the work from run time to compile time. In general, this is a good thing.
I just don't think that the savings are all that important, and there are things that cannot be known until run time. The compiler cannot do everything.
Most modern microprocessors (and I guess the bigger ones too) have fetchers that read in instructions, break them down, and feed the pipelines. All the complexity of CISC goes away at this point. Instead of one instruction being pushed into a pipeline, two or more instructions get pushed (hopefully into different pipelines). This here is the wonder of the superscalar concept.
The fetcher gets a bit more hairy, as does some of the speculative branch handling (more things to invalidate), but my goal wasn't to simplify the hardware. Anyway, the hardware already has some of this, so it isn't like we're adding anything new.
Conceptually, I prefer the idea of pushing as much of the info down as possible (CISC) so that the lower bits has a larger view of what is going on. Think of "peephole optimization" in a compiler: the larger the peephole, the better the optimization (within reason).
Compilers are handicapped that they often compile to the lowest common denominator. In the IA32 world, the instruction set hasn't changed too much in years, so many people forget that not all CPUs in the same family are identical. Think about the older PowerPCs: 601, 602, and 604 (I'm showing my age, eh?). The 601 was a hybred, so it had some odd opcodes. The compiler had options to generate for one of the CPU types, or target the lowest common denominator. Guess which one most applications used? Even if you stuck to the 603 and 604, life wasn't much better. I think that they had identical instruction sets (it has been a long time), but I'm sure that they had different pipeline geometries. It is not possible for a compiler to generate code that is optimal for both. That is why code that is optimal for a 386 is not optimal for a PII, even though the instruction set is the same. Details matter. The compiler can't know them all.
As long as I'm rambling, this is the main reason I've gone anti-VLIW. If a merceed comes out with 4 actions per instruction, and a mckinley has 6 actions, how is the poor compiler going to optimize for both?
Let the compiler parse, hack, and optimize all it can, but there are somethings that cannot be known before run time, and we need to let the CPU handle them. I think that CISC helps this out.
Have I made my arguement any clearer this time?
- doug
PS: I cut my teeth back on the old M68000 family and it is still my favorite instruction set. I worked with PowerPC 403s for a while, which is where my coding RISC assembler background lies.
PS2: To be honest, as long as the CPU isn't little-endian, I'm not going to get worked up over it. This is all just quibling over details.
Re:RISC vs CISC is meaningless (Score:1)
I am fully aware that newer IA32 CPUs (PII, K7, etc) have "RISC cores", and I think that this is the way to go (although I dislike the IA32 instruction set). RISC makes no software person's life easier, and CISC does. I don't care so much about the EEs doing the hardware, as I'm not one (yes, I'm callous).
You are right that "RISC is supposed to make the hardware solve the same problem in a smaller amount of time than a CISC design in the same process with the same constraints (cost, power..) and the same amount of development money." The gain in RISC though wasn't in the instruction set per-se, it was in letting the different parts of the CPU work in parallel. That can be done in CISC too, although some of the most complicated instructions may need to go away. C'est la vie.
Newer CPUs require that all instructions are the same length. This is required for RISC, but not for older CISC machines. Most modern CPUs (hybreds, the whole lot) have this restriction to simplify the fetcher's job. This is a fair tradeoff although some funky instructions go away, but it is worth it.
I like the observation that "Some say it's because pre-RISC cpus were designed for assembler programmers and RISC CPUs are designed for compilers." This is more that most programmers would go crazy coding big stuff (whole applications) in RISC. Compiler folks are already crazy, so it is no loss. Fortunatly, most of us code in C or something else, so this is a moot point.
As for your arguement "The flaw with this example is that you assume that we won't use C for some time", which is true in that case. Lets try a different example: think about a stack pointer. It is a common tactic to
move value => (--stackpointer)
There are three primitive instructions here: 1)change the stackpointer register, 2)get value from memory, and 3) store value in memory. It is quite possible that "value" will be used again in the near future, so saving it in a register is useless. Obviously as registers become more common, there is less motiviation to conserve them, so maybe this isn't important.
- doug
How important is platform architechture (Score:3)
Re:RISCy statement (Score:1)
John
Re:64-bit NOW! (Score:1)
This is great for BeOS! (Score:1)
-o
If the BeOS doesn't work on these things.... (Score:1)
Re:RISCy statement (Score:1)
The main difference now is that RISC chips generally won't let you do things like add from a location in memory without loading the value at that location first and x86 will still let you do that. If you want to write fast x86 code you will write it like you would write RISC code. The philosophies are very different still. Intel has just been good at adopting the ideas that IBM, Mips and DEC put out first.
What RISC and CISC don't mean is a way to measure performance, that's why the marketeers use it but most users and probably even a lot of programmers don't know and don't need to know the differences.
More good news (Score:1)
IBM and Motorola are in a curious position, they have developed a good modern highend processor but because of the cost factors associated with PCs today they are having trouble pushing as many as they'd like. Likewise both have invested enough and depend enough on the architecture that they can't kill it. Free specs and cheap mobos only bodes well.
Look at the netdriver, very cool, very sexy, very expensive for what you get. If they could cut a few hundred dollars off the price, you'd have a top notch internet appliance, a serious iMac competitor. I think the rationale from IBM could be one of two things, it could be good will, they had something they didn't need so they went public with it or it could also be that they think that if PPC mobos drop in cost enough they believe that they can compete with Intel and AMD on a manufacturing cost basis and as PCs continue to drop in cost the freeness of linux will begin to play a huge factor.
If you're building linux based internet appliances, hardware cost is you only problem. They are already committed to making more and better processors and will be for some time to come. It's a good move on their part and I think the community will benefit too. If I could buy a PowerPC chip and mobo for just a little more than an Intel, I'd probably do it.
Re:Only if Linus comes down to earth again. (Score:1)
Re:How important is platform architechture (Score:1)
The real question is how well written the original software is. Odds are if you can port it from x86 to sun4, you can port software to ppc with a simple recompile.
Re:fracturing Linux... (Score:1)
Perl is a good choice. So is tcl/tk, or any other cross-platform scripting language.
If you want cross-platform binaries (for those closed-source addicts), then use Java.
If AMD invested 2 Million... (Score:2)
Then, maybe AMD would really blow the doors off of Intel ;-) And for a cost much less than $200 million.
Compiler Influances (Score:3)
Intel's own compiler for the Pentiums is very good, but GCC is also great for x86, so it's popular. The commercial DEC (rr... Compaq) compiler really rocks on Alphas, but gcc isn't nearly as good for Alphas as the commercial compilers. So, Linux/Alpha isn't nearly as popular as you would expect it to be (give shear preformance factors of the CPU are masked by the results of the compilers).
I have no doubt in my mind Linux will run on almost any platform, the Linux community is very very active in getting the OS ported to new hardware.
I have doubts that PPC will become popular. If Motorola or IBM puts some money, work, and support into GCC, then the G3's will really rock in Linux. If they don't, it'll just be "another" platform that Linux runs on, but nobody really uses (much like Alpha is now). Before you consider this a flame, check benchmarks of Comercial C and Fortran compilers for Alphas and benchmarks for gcc on Alpha. And, then notice that there are a lot of people who would consider Linux, but end up buying a commercial OS and compiler for thier Alpha insted.
actually (Score:2)
--
Only if Linus comes down to earth again. (Score:1)
The so-called stable 2.2 kernel that claims to support PPC won't even compile on PPC! And that's not because there are no patches, it's because Linus refuses to include the patches before releasing a new kernel. He even intentionally breaks support for some platforms, as has happened in the 2.2.3 kernel! If you want to get a kernel that actually compiles, you'll have to find the (undocumented) directory on vger and check out the tree with an (undocumented) CVS tag. Just forget about going to ftp.kernel.org, it won't work.
Here's a hint: CVS _does_ work. Delegating work to other people _does_ work too. Do it for the main kernel. Now.
Linux on PPC has a great future, but not as long as some bonehead on the top is blocking it.
Re:Only if Linus comes down to earth again. (Score:1)
More PPC sales means more Mac sales (Score:2)
A customer buys 100 PPC boxes to run 100 web servers. Now he needs a client desktop. He'll consider a Mac more than before because it's also PPC - "Just in case those Macs don't work out so well, I can turn them in Linux boxes like the ones I already have."
Timur Tabi
Remove "nospam_" from email address
My answer, probably not (Score:1)
Longtrail + 604e@225 was supposed to be 450 USD (Score:1)
Re:How important is platform architechture (Score:2)
Linux already does that, and does it very well. I can compile virtually everything on my Sparc Linux box just as easily as I can on my Intel ones. The only exceptions are the few dolts that assume Linux == x86, and do things like include x86 assembler for a few routines ("for performance"). That's all well and good, but it makes your app gratuitously non-portable, when it needn't be. autoconf should be able to detect the platforms for which you can substitute fast hand-crafted assembler for slower but functionally identical C routines. That gives you proper portability with performance benefits on certain platforms. Either way, 99% of apps that use autoconf just compile straight out of the box on all my Linux platforms.
Re:Linux will keep processor makers on their toes. (Score:2)
Yep, couldn't agree more. Virtually everything else out there is superior in terms of design, build quality, etc., but when it comes down to it, market pressures have forced PC prices down so much that everything else is just not good enough value. I love my Sparc to bits -- PCs don't even come close to the simplicity and elegance of its design (why, oh why, haven't SCA drives become commonplace in the PC world?). However, your average punter isn't going to spend money on a decent RISC machine to get the same performance as a PC costing half as much, no matter how good the build quality. At the high end, pricing is closer to parity, but that's mostly due to Intel's extertionate pricing of Xeon's so they match equivalent Sparc / Alpha / MIPS offerings.
If there was a cheap PPC option, I'd almost certainly go for it. That said, I'd still have to keep my x86 boxen to run those binary apps that don't yet have an open source equivalent of sufficient quality.
OT: Did ANYONE notice the item about ATI? (Score:1)
Being the paranoid sort, I asked whether ATI would permit the resulting drivers, whose development would be based on detailed -- and presumably NDA'd -- information, to be open source. The answer was a definite "yes." '
I don't know about you all, but it's finally happening- all of the 3D vendors are getting clues by the bushel load and they're making drivers happen.
Re:Eh? (Score:1)
Choice seldom makes it worse for those doing the choosing (Linux folk in this case). Choice does make it worse for those being chosen. The PPC would be in much better shape if it were the only game in town.
In this case I think it is hard to not choose the alpha if you want maximum single CPU speed, or the x86 for minimum price.
I'm not sure what gets you the best bang for the buck. I expect the ARM gets the best bang per milliwatt (even the new 600Mhz intel ARMs run fairly cool).
I honestly don't have a clue what the PPC is best at other then running PPC code!
I don't see why you would think that. My original post disn't espouse anything (except maybe cheep SS7 motherboards). This post has a bit more espousing in it. However you should note I never espoused anything merely because it was popular. I did say PCs were cheep, and it would be hard to beat them on price. However that is (a) true, and (b) not a popularity argument.
Thank you. I may, of corse, be over excited about the iBook merely because of my fondness for 802.11.
Eh? (Score:3)
Frankly I don't see why a cheap PPC motherboard is going to make a huge diffrence. PC motherboards are quite cheep, under $100 for a Super7 motherboard. So if the PPC is going to compete in price it has a long road to hoe. A free design doesn't mean free motherboards, in fact the free design might not be as cheep to make as some of the more mature PC motherboards!
The fact that Linux is more or less processer agnostic just makes it worse, after all why go for a PPC rather then an alpha just because there are vague roumors that the Alpha will gasp it's last any year now? I mean if switching to a new CPU is so easy, why not use the Alpha until it actually gasps it's last? (assuming of corse that the Alpha is faster -- with the SPECmarks seem to say, and cheap enough)
The only real argument I could see for using the PPC is if it (the PPC) actually made it into nice cheep machines, like maybe the portables (they seem relitavly inexpensave for what you get -- but I havn't looked at PC portable prices, so i may be in for a shock). Actually that isn't the only argument. It would be intresting to see Linux on one of the big PowerAS machines...nicer still to see it hosted under VM on a 390 (but that's not a PowerPC).
Woo-hoo! IBM, Motorola, Loki!! Woo! Yeah! Woo! :) (Score:1)
I hope this blossoms, and we have a REAL price and performance war between x86 and PowerPC so we'll all benefit from better execution not just cheap hot running Intel processors.
I really don't think Compaq will pull it off with Alpha Linux... their leaders needed vision on this a LONG time ago, and there's too much internal bickering and backstabbing. SGI was SMART when they ditched their NT division... when you make and sell an OS or OS' who wants departments with loyalties divided with the competition?
Some folks I know did a "R.I.P" on SGI when they cutoff their NT division, but I think this was smart.
You can say you think I'm smoking crack, but I think Jobs has already laid down some groundwork for Apple to become a Linux company whenever it becomes necessary. (If you doubt this is possible, think about how difficult it would be for OS X Server/Consumer... not at all, and it would be one giant fsck you to Bill Gates in the history books...
Anyways, more CPU support in Linux is better. I agree Motorola and IBM better commit some resources to GCC if they want to be taken seriously - it's a relatively small problem to solve.
I'm still completely blown away that Loki's supporting all the Linux games on PowerPC. This is something I hoped for and banked on happening about 1 or 2 years after Linux became viable for Commercial games... not MONTHS as it has turned out. Linux is looking more and more unstopable.
Re:Does anyone know where these specs are? (Score:1)
It's the binary-only stuff. (Score:1)
This sw is probably closed source (Score:1)
I love my G3 (Score:1)
-Ben
Re:How important is platform architechture (Score:1)
I'm running Linux on AXP and SPARC for quite some time now.
They both have a smaller user (and developer) base than
Intel platforms and it shows! Network code is far from being
as stable as on Intel. A number of applications won't compile
on AXP at all because they aren't 64bit clean. Others will
compile and even run but crash after a short while. (Thank
goodness, that isn't true for most of the KDE apps.)
If PPC gains a really large user base it will mean that
Linux/PPC will become more stable and reliable.
If only that were true for AXP!
Uwe
why would macworld publish this? (Score:2)
ibm's move isnt going to help APPLE any is it?
i mean, seriously, aesthetic considerations aside, people tend to like apple hardware but hate the (technically speaking, now...) OS. if this does spawn a clone war apple could be skrewed.
what if individual companies try to sell g3 boxes with more features and better price points than apple? apple cant very well revoke licensing or buy out competition THIS time...
who here would buy a g3, considering the architecture and processor power, if you didnt have to subsidize apple's os which youre not going to use anyway?
i know i would.
unc_
Could BeOS gain anything from this? (Score:1)
Be's explanation of why BeOS doesn't run on G3s. [be.com]
-G.
Re:I love my G3 (Score:1)
Linus tries to keep the Linus Kernel as fair on all platforms as possible -- but unforently he is a very busy man, and sometimes he loses / messes up PowerPC patches, just like he sometimes messes up Alpha or Sparc ports (although he typically doesn't ship stable broken i386 versions
Personally, I don't find it acceptable to be shipping stable production kernels that are broken -- stable to me, means it works up to the promise -- and is not unstable (if you want a broken kernel get 2.3.x, you)!
2.2.0 had support for the PowerPC--but recently Linus (and the powerpc kernel deveopers didn't get there patches in at time) had some issues. This is unaccepatible for a stable kernel -- but I guess Linus doesn't think it's important enought to make sure a Kernel 100% stable before shipping it marked stable.
One more thing to note: Linus's tree might be good for some -- but it's highly recommended that you get your Platform's stable spefic kernel (such as vger-ppc 2.2.x or vger-alpha 2.2.x).
So for the last time, shipping defective / broken code in a 'stable' product is just unacceptable.
Re:This guy is an idiot (Score:2)
Apple dumped the 604e from there line, because Apple wanted to make the PowerMacintosh line, cheap, simple and easy. So they all use the same processor (a consumer on), logic boards that are very much similar, and cheap PC RAM, that standard on all current machines they sell.
Apple won't go G4 (likely) untill they discontinue making all of there G3 systems -- and that may be a while.
This helped then reduce inventory, and become lean and mean -- no extra baggage.
Of course this pissed off high-end PowerMac customers -- they are either too slow or lack to many PCI slots to be usefully. But it made iMac possible, and cheap for Apple -- but it came at a cost.
PowerPC would be useful in the portible market -- except for one big problem -- there is no CHRP portables ever made -- they are all big desktop machines, and no portibles. Maybe somebody can design a portible machine....
Dominant for low to midrange server-ettes - maybe (Score:1)
New iBook needs no cooling fan... (Score:1)
Ahhhh, LinuxPPC would smoke on this box...
http://www.mosr.com/terms.html#UMA
http://www.appleinsider.com/macworld/mwny99/ibo
http://www.ogrady.com/models/p1.asp
Alternative platforms are good for security (Score:1)
Cheers,
Ben
It's bang for the buck that counts here (Score:1)
That's the reason to be interested in PPCs. Low prices come from volume production, and Alphas just ain't got it. PPCs have a start in that direction with Macs, and it's possible that adding Linux would boost production enough to keep the price low. Since PPCs (like Alphas) make better use of silicon die space, they have an inherent advantage over any x86. It just takes volume production to realize that advantage.
--
Re:not the only one, but.. (Score:1)
Re:Apple Powerbook G3's (Score:1)
I'm not too concerned about the lack of a three button mouse. I think that the Linux environment I'll be using is configurable enough to make up for that. Pasting text is one of the few things I'd need to adapt to. Can't say I use the right and middle buttons on this Thinkpad 770 for much more. I might as well be pressing a keyboard modifier for my speed with any other operation.
I am perhaps fortunate in that I don't have an investment in architecture-dependant binary software on Linux. Since I don't plan to run either office-typical suites or big relational databases on my laptop, there's no problem there. I'm happy to see that there's both an x86 and PPC Linux version of Xing's MP3 encoder, so I'll be able to build up my MP3 library as I use LinuxPPC, using modern, licensed codecs.
Hardware-wise, the G3 Powerbook kit is ahead of the game WRT most x86 laptops. Anything that comes close is at least as expensive. The real plus is that when it comes to running mainstream proprietry applications, the MacOS is a pleasure to use, and Windows is a chore. If, like me, your use for a "toy OS" is to run media-creation software rather than play games, the MacOS and hence PowerPC is a better place to be.
Finally, it's *way* cooler to be running a hacked-togther LinuxPPC on a G3 'book than a stock RedHat on a Thinkpad.
USB support (Score:3)
This having been said, my only problem with the PPC architechture is that so many darn PPC machines still use one-button mouses...
What good is in G4 (Score:1)
Re:What good is in G4 (Score:1)
My dream would be something like GNU Scientific Library, but in C++ (it is very contrproductive to write scientific analysis software in C, sorry all C fans..) and with MMX/Altivec enhancements available...
Re:What good is in G4 (Score:1)
That's true. And I do not write assembly either. But I do write a lot of numerical analysis code for my research and would like to know if any generic libraries using SIMD are available. Especially for C++ STL - so I can write getChecksum(vector& Data) and it uses MMX2/Altivec or whatever to calculate that faster then the build in alogorythms or my hand code.
Thank you for pointing at my ignorance - but you did not answer my question
Re:What good is in G4 (Score:1)
Re:Only if Linus comes down to earth again. (Score:1)
The preferred platform for Linux will always be... (Score:1)
Right now, computers based on Intel and x86 compatible CPUs outnumber everything else. And thus, they are the preferred platform for Linux.
As soon as another platform emerges, I'm sure that some folks will port the compiler tools for that hardware, others will work on the kernel, yet another group of hackers will port XFree and in the end, it will be fully supported -- if only enough people feel the urge to use this other platform with Linux.
And because of this, the question "which is the preferred platform for Linux?" is pointless: It's the platform you already own.
Re:And the answer is: Nope. (Score:2)
Well, it might mean that gcc has better optimization on x86/Merced than it does on PPC.
--
Re:And the answer is: Nope. (Score:2)
Well someone mentioned above that Apple has been folding PPC changes into gcc.
--
Apple Powerbook G3's (Score:3)
This cost 2300$ retail and for the hardware it appears to be an absolutely great deal. I have a few other new PC laptops here from compaq and toshiba. In the same price range they don't have built-in ethernet, they don't have SCSI, and they don't have as much video RAM. They are also Celeron or AMD K6-x chips, not P2-3's. In my subjective opinion the G3 laptop is cooler looking as well.
So.. before you go and bash apple powerbooks, check out the specs, pricing, and use one for a week. All my other computers here are PCs but you have to give credit where credit is due. Apple's G3 powerbooks are real contenders, even leaving out the OS.
PS: Compaq prices their consumer laptops very low, but who would want a "retail" button with a shopping cart icon right next to the trackpad, even if the rest of the specs rock? GRRR.
Outland Traveller - new and laptop enabled!
Re:The main processor for Linux... (Score:1)
The newbies have always outnumbered us. Exponential growth is like that.
What might be different is that we're getting users less interested in learning how things work.
the specs don't appear to be publicly available (Score:1)
I did find schematics for the earlier reference designs that have been pointed out as counter-hype:
1995 December dual 604 design [ibm.com]
more recent 'spruce' 6xx/7xx reference board [ibm.com] (uniprocessor, based on the CPC700 hostbridge)
There's no license on either of these, though, so it may be the 'free' part that's new.
To counterflame:
Hardware designs can benifit from open development in all the same ways that software can--faster development, better designs through the pooling of resources and peer-review. Our community can benefit in the same way as well--no one can take control of your computer away from you. How dare you flame someone for wanting to hack!?
Yes, prototyping hardware is expensive, but a large part of that is because fabs are designed for mass production runs. There's a fixed cost associated with setting up a particular design, whether your making a couple of prototypes or 10,000 units. I bet we could design a fab for low-volume production that would be the other way around. And no, it's not cheaper, but it's definitely possible to build motherboards in your garage.
What really bothers me here are the off hand dismissal of a call for openness. Are you trolling? Remember, amateurs can't write production quality software. Something as serious as an application suite (I won't even speak of an operating system) can only be designed by a qualified team of professionals, and their work costs $100 per line of code. What are you going to do, hire some college kid to do it?
slashdot to the rescue (Score:1)
Thanks for the pointer!
processor diversity is important (Score:2)
All the processor dependencies that have crept into Windows and its software architecture (ActiveX, drivers, etc.) are one of the biggest problems Windows is facing, and it is good if Linux can avoid falling into the same trap.
Linux/Alpha held back by GCC backend. (Score:2)
But Intel understands one thing that DEC doesn't seem to: providing a free, open source C/C++ compiler is essential for success with free software. In effect, that approach recoups the cost of compiler backend development from the people who buy the chip; charging for the compiler puts the burden on the software developers. But since developers of free software already donate their time, charging them for the compiler doesn't make a lot of sense.
In different words, at least in the open source world, the GNU C/C++ compiler is essentially part of the processor itself, and if GNU C/C++ doesn't perform well, then it doesn't matter much how fast the processor is with some proprietary compiler.
Unless DEC sponsors work for improving the gcc backend for Alpha to be competitive with their own proprietary compiler, my guess is that the 64-bit Intel chips are going to win.
Re:How important is platform architechture (Score:1)
Or you can still use Intel. If LinuxPPC is as close to a linux distro as you need, I don't see any reason NOT to be using a PPC, especially Since Apple is on the brink of releasing G4 boxes.
Re:fracturing Linux... (Score:1)
That's the big deal anyhow.. I have an Alpha. My biggest deal is not linux at all(stable as hell). My problem is getting drivers (3D hardware acceleration comes to mind.)
I've noticed that alot of the cross-platformism is starting to dwindle a bit. One of the reasons that Linux became so popular was the joy of hackers digging into hardware. (Atleast for me).
Pan
Alpha, PPC, and MIPS all in the same boat.. (Score:1)
Hardware Peripheral manufacturers (Certain 3D card makers like 3dfx) refuse to support the alpha. ATI hasn't been that great (Maybe changing soon?). There are just loads of peripherals.
Mr Alpha and Ms PPC had better let everyone know that we need specs. As companies come in and dump their binaries for x86 out there, we have to let them know that that's not good enough. You either support LINUX or don't. I'm not saying that companies have to give out their family jewels, but they should be willing to allow people to use whatever platform they want.
They don't realize that a good programmer can build cross platform code. Bad ones don't. Look at LinuxPPC and AlphaLinux. 99.9% of programs compile out of the rpm. (The only exceptions being ones that are very hardware dependend and include lots of x86 asm "speedup code")
We should have a platform independent stance. As more people and newbies move into Linux, they arn't going to care what platform Linux runs on. This is a good thing.
By keeping Linux Plaform Independant, we arn't tied to the death of Intel by Transmeta.. or Elbrus, or whatever.
Pana
Re:Bang for the buck.. (Score:1)
Re:the article mentions Altivec... (Score:1)
If my understanding of AltiVec is correct, it (AltiVec) by itself won't do much. Unless Linux (or MacOS, or DarwinOS, etc) takes advantage of it.
Re:Eh? (Score:1)
Erik
Has it ever occurred to you that God might be a committee?
Yes there was... (Score:1)
Erik
Has it ever occurred to you that God might be a committee?
Re:RISC vs CISC is meaningless (Score:1)
Claim: The last general-purpose-designed-for-performance CPU that's internally RISC like was the DEC/CPQ Alpha 21164, today the core of every such CPU is very much like dataflow!
The "breakthrough" of the RISC design wasn't coming up with simplier instructions, it was breaking the CPU into independant subsystems that could work in parallel. Doing an integer operation at the same time as doing a floating point one, while loading data from the bus/memory means 3 things are going on at once. A "simple" instruction set means that it is more obvious where the pieces are, but is fundamentally irrelevant.
A few things to consider;
Everybody does not agree 'why RISC is better than CISC, and why it came when it came'.
Some say that it's all in the title of a book "Computer Architecture A Quantitative Approach". i.e. that pre-RISC designers had a flawed design philosophy.
Some say that it was because you could finally get enough transistors on a single die to make a pipelined non microcoded general purpose CPU in a single die.
Some say it's because pre-RISC cpus were designed for assembler programmers and RISC CPUs are designed for compilers.
And: What makes RISC RISC?
Is it the load-store architecture (must load data into registers before manipulating it)?
Is it the uniform insn length?
Is it the design philosophy?
IIRC the first RISCs (the first CPUs that were called RISC) didn't do multi-issue of insns in the same cycle and they most definitely were not out-of-order..
Oh, well - mostly pointless for anyone but historians and markerers
When programming in assembly (the only time RISC/CISC is visible to a programmer), you want it to be as CSICy as possible, it makes your life easier. Think about
C = A + B
In an old assembler that is one instruction. Simple, easy to read (for assembly). In a load-and-store system that is
load A to r1
load B to r2
add r1 to r2
store r2 to C
4 vs 1. You tell me which one you want to hand code in.
The flaw with this example is that you assume that we won't use C for some time - usually C will get used again (maybe the only reason it ever was in main memory was that the CISC where the asm was written was register starved)? But yes - RISC systems generally do more insns than CISC systems - just not four times as many
Also, RISC pushes more of the work to the software, which is fundamentally slower than hardware. Lets do as much stuff as low as possible so our systems run faster. Adding layers sucks.
I assume this is irony? Software in-and-of itself has no 'speed' - only when executed on hardware does the software have 'speed' - and RISC is supposed to make the hardware solve the same problem in a smaller amount of time than a CISC design in the same process with the same constraints (cost, power..) and the same amount of development money...
Imagine a future with CISC instructions with pipelined cores. How one gets translated to the other is meaningless to programmers although an interesting research topic for the hardware folks. Think of it as a library. You just care what the interface is (the opcodes), let the library designer (EEs) handle the details of getting that interface to work well. Maybe some of the ease of the monster CISC stuff needs to go away to help out the core (trade-offs are part of an engineer's life), but as a rule CISC is better for programmers. Just lest RISC vs CISC be forgotten. We need bits and pieces of each one.
Must be heavy irony, because thats what we have today - the P6 (used in PPro, PII and PIII), K6 and K7 cores are multi-issue super-pipelines ones, they happen to be implemented as translators to 'internally RISC' that's really a kind of dataflow (just like all current RISC CPUs)..
Erik
Has it ever occurred to you that God might be a committee?
Re:Because of the 1.1 GHz G4 chip (Score:1)
The IBM PPC demo was really room-temp but it was not a complete CPU (no fpu etc.) and was just a 'technology demo', but it also used a older process technology (It was around when IBM announced copper interconnects but the 1GHz demo was in aluminium)
Erik
Has it ever occurred to you that God might be a committee?
Re:Because of the 1.1 GHz G4 chip (Score:1)
Has it ever occurred to you that God might be a committee?
64-bit PPC exists (Score:1)
apparently there is a 64-bit PPC. IBM uses it in some of their workstations. lookit.
http://www.austin.ibm.com/resource/technology/n
"BM's NorthStar superscalar RISC microprocessor integrates high-bandwith and short pipe depth with low latency and zero cycle branch mispredict penalty into a fully scalable 64-bit PowerPC-compatible symmetric multiprocessor (SMP) implementation. Based on PowerPC architecture, the first in the Star Series of microprocessors, the NorthStar processor contains the fundamental design features used in the newly available RS/6000 and AS/400 server systems targeted at leading edge performance in commercial applications as characterized by such industry standard benchmarks as TPC-C, SAP, Lotus Notes and SpecWeb. "
i'd imagine that you ought to be able to easily cobble together something based on this chip. it might not be able to run MacOS, but linux is the point of this thread anyway.
- mcc-baka
who needs sheepshaver? http://www.ibrium.se/linux/mac_on_linux.html
What goes around.... (Score:1)
I myself run both flavors of machines, and prefer the LinuxPPC R4 (gotta get me the 1999 issue soon!)on my Apple 6500 PPC 250 over RH, Caldera or S.U.S.E. on my Intel boxes. IBMs latest action will only make the PPC boxes more mainstream, and therefore better supported. I can't wait till next year.
Re:Bang for the buck.. (Score:1)
So far, it has been so-so, mainly because Apple is the only company that sells PPC computers for desktop users, and Apple stuff is overpriced. On the other hand, a few years ago (before Apple backstabbed the Mac cloners) the PPC perf/cost was pretty good -- better than x86. The hope provided by IBM's latest announcement is that the cost of making PPC systems will go down. This would increase the perf/cost ratio again, competitive with or exceeding x86.
Yes, assuming someone writes a driver for that card and the OS that you're going to run.
One "issue" (just off the top of my head) is that some cards have some software burned into ROMs, and that software is probably written for 8086 and makes calls to an IBM PC BIOS (and probably extends that BIOS as well). While someone will write drivers for your cards, you might not be able to press Ctrl-A during boot up to reconfigure your Adaptec SCSI card's settings, for example. You'de probably be able to run a program like that after booting, but that sounds like it could get chicken-and-egg-ish.
I think there was supposed to be some kind of fix for this issue back in the CHRP days, which involved putting processor-neutral code on PCI cards. I dunno much about it, though. (Is this what "Open Firmware" was?)
Re:Could BeOS gain anything from this? (Score:1)
Sorry to pick nits, but this is deceptive. BeOS' compatability problems are with recent Macs (which just happen to be the only machines running G3s), not the G3s themselves. Building a G3-based machine that works with BeOS shouldn't be a problem.
Re:And the answer is: Nope. (Score:2)
And is Intel going to attach a string to this money, saying that Red Hat must withhold source code? Are they going to encourage people to code in assembly language? Not likely. All those Intel dollars being spent on Open Source will be very useful to the PPC users who don't mind typing "make".
not the only one, but.. (Score:2)
x86 is not RISC. The point that i get from this is not that PowerPC linux boxen will overtake x86, just that the Merced will be all too expensive. Opening the door for a more afordable alternative, which alpha has proven not to be. The point is that PowerPC with the help of IBM's proposal will make powerful RISC systems that run linux which are more affordable than the Merced alternative. If the archetecure bound to it is proven to be superior, its all up to the programmers after that.
Long Live PowerPC.
And the answer is: Nope. (Score:2)
See the Red Hat Wealth Monitor [prosthetic-monkey.com]
Re:Apple Powerbook G3's (Score:1)
I'd buy a PowerBook and never look back if they would make one with three buttons like the ThinkPads.
Anyone know if MacOS X is going to continue the one-button tradition?
Re:Intel for a long time (Score:1)
Of course nobody's going to throw away their old boxes, whats your point.
It doesn't matter that the market may suddenly swell a little with new PPCs that are less expensive than Apple boxes; it still entails SPENDING money instead of using what we already have.
Well, duh. New x86 computers cost money too.
The whole point is that if you're building a new Linux box you can go with an affordable RISC platform instead of just another Intel machine.
If you're talking about recycling old hardware, you should be able to use old hard drives without problems, and maybe your RAM and some PCI cards.
Re:Motorola's Free PPC Motherboard...remember?? (Score:1)
Re:The preferred platform for Linux will always be (Score:1)
Uh, no, that would the the one with the most market share.
Right now, computers based on Intel and x86 compatible CPUs outnumber everything else. And thus, they are the preferred platform for Linux.
Right, like Windows is the worlds most "prefered" operating system, just because its running on more machines than anything else.
And because of this, the question "which is the preferred platform for Linux?" is pointless: It's the platform you already own.
Just because you're running Linux on a 486 doesn't mean that the 486 is your prefered processor. I'm running a dual celeron but would prefer to have quad K7's.
Actually...... (Score:1)
It's entirely possible, because both companies sub-licensed the MacOS to other companies, and Motorola even built their own line of SuperMac clones. Understandably, this pissed of some execs (mainly Moto's CEO) and is one of the reasons why PPC development is lagging behind.
If anything, Moto and IBM will push for these boards to get back as Jobs for stabbing them in the back. Hope they do it.
The use of processor power (Score:2)
Well, the correct answer is the usual one: it depends. The location of the bottleneck is highly dependent of what exactly you are running. Some of the processes I run are I/O-limited and having faster hard drives would speed them up. Others are bandwidth-limited (yes, on a LAN) and a gigabit Ethernet would help. But most of my stuff (guaranteed to be untypical) is actually CPU-limited - and I am running on a dual Sun Ultra 60.
So, yes, I understand the importance of the bus, and DMA, and AGP and all the other TLAs. But for me, at least, processor speed is more important right now.
P.S. For example, in the FPS (Quake, etc.) gaming community there is a very well understood distinction between being CPU-limited and fillrate-limited. Depending upon your specific hardware, any of these can be your problem.
Kaa
Laptop Market (Score:2)
You don't really need 64-bit. (Score:2)
The PPC design has 64-bit extensions that would be more than adequate for your 64-bit needs (file-systems or whatever). Lets face it, the majority of your integer work is perfectly happy in 32-bit, and when you do need real double-precision floating point values, the PPC is happy to oblige you in native 64-bit floating point arithmetic.
More importantly, the amount of power the average user needs is not increasing quite so rapidly anymore, especially for those not dependent on Micros~1 products (which have allegedly been deliberately bloated to force users to buy newer hardware). Does Uncle Joe need a GHz machine to browse the web? Hell, at the speed he moves the mouse we could have gotten him a 486 and he wouldn't have noticed the difference.
And while it may not be the most powerful architecture on the block, it certainly kicks some ass in the 32-bit world (and I've studied the architecture). Given the alternatives, I think that it definitely deserves a "best in its class" type award.
The more important question to ask is (and people have been asking this), will it be affordable? If it is, it will succeed and everybody will be happy (myself included). I think that it's a beautiful architecture, and would love to be able to buy an affordable computer based on it.
Alpha (Score:3)
I think they missed an important point here. The PPC is wonderful arch., but it isn't going to be long before the industry starts the Big Move to 64-bit (Merced will probably be the catalyst, right before it bombs[0]). Unless IBM is planning a G5 based on the PPC620, this will leave them behind.[1]
Of course, if the price/performance favors the G4/5 enough (say, dual 800Mhz G4's for the price of a 1GHz Alpha) then it may get ahead, but otherwise any victory will be short-lived.
[0]: I still think that IA-64 is a plot by HP to kill Intel. Instead of trying to compete with them, HP offered to help design the new arch., then came up with something so horrible that there is no way it can succeed.
[1]: Before anyone flames me for implying that 64-bit is always superior, think about this: Once the industry begins the move to 64, the Alpha, Sparc, etc will all be there waiting. If Merced bombs, the chip most likely replace it is the Alpha, which blows away a PPC. Demand == lower prices.
Re:why would macworld publish this? (Score:2)
In addition, many Mac users feel like kindred spirits to Linux users. The operating systems may be vastly different, but the rejection of Windows and belief that the OS does matter makes the Mac community pull for the Linux community.
Finally, if Linux became popular on PPC, the market for PPC hardware components would decrease as the market size increase (good old economies of scale).
CISC is bad mmmkay (Score:2)
My understanding of Merced is that much of its speed comes from the compiler building in hints for out of order execution. In which case 1) Assembler by hand will be harder anyway and 2)Old software will need to be recompiled to get the most out of the chip.
The first point doesn't bother me. I did a bunch of MIPS assembler in school. That was enough. Compilers these days do a pretty good job. Let them!
The second point doesn't really matter for linux because we've got the source for damn near everything. It's not so easy for Windows where there's tons of assembler everywhere and updates will cost the end user...
Re:This guy is an idiot (Score:2)
It's well known this isn't the G3's strength.
Not today, but soon. The G4 has great FP *and* altivec to boot.
Also, as a previous poster pointed out, Altivec is Motorola's baby, so I have to wonder if it'll be supported by IBM's spec.
IBM has agreed to support Altivec.
Hopefully the PPC will come of age in the portable market where power consumption matters...