Khronos Releases OpenCL Spec 115
kpesler writes "Today, the Khronos Group released the OpenCL API specification (which we discussed earlier this year). It provides an open API for executing general-purpose code kernels on GPUs — so-called GPGPU functionality. Initially bolstered by Apple, the API garnered the support of major players including NVIDIA, AMD/ATI, and Intel. Motivated by inclusion in OS X Snow Leopard, the spec was completed in record time — about half a year from the formation of the group to the ratified spec."
Re:yeah... (Score:4, Funny)
Yes, but you get 2**256 very tiny virtual consoles on screen, each with only 128bits of ram. On the up side, every console can be at a slightly different angle, with different specularity.
Dear moderator... (Score:1, Offtopic)
I don't think redundant means what you think it means ;)
If I had mod points (Score:1, Offtopic)
what does it DO? (Score:4, Interesting)
is this simply a spec that people expect ati and nvidia to conform to? or is this another api outside of CUDA and CAL, that wraps the two up so that a single api can execute code on all GPGPU's?
Re:what does it DO? (Score:5, Informative)
Re:what does it DO? (Score:4, Insightful)
If the past few years is any indication, it works much better on GPUs than on CPUs.
Re:what does it DO? (Score:5, Interesting)
Re: (Score:2)
Re:what does it DO? (Score:4, Informative)
CUDA is already doing great things in molecular dynamics, which bears some similarity to FEA:
HOOMD Benchmarks [ameslab.gov]
A single 8800 GTX reaching 75% of the performance of a 32 node cluster isn't bad. I imagine the GTX 280 would easily beat the cluster.
Re:what does it DO? (Score:4, Insightful)
The two major issues to be solved with that are that you need double-precision hardware (I can't remember if the Nvidia 9000 series supports that or not) and, more importantly, you need to write GPU algorithms for solving sparse matrices.
Re: (Score:1, Informative)
I about 99% sure that the 8800/9800 series are single precision, and gt260/280 have double precision. I'd guess everything beyond 260/280 will have double precision, but you never know.
Re: (Score:1, Informative)
Oh, and you need floating point exceptions, not silent over/underflow.
Otherwise you get nonsense out the other end and have to start over.
I've yet to see a graphics card that's fully IEEE-compliant.
Re: (Score:2)
If Nvidia or ATI really wants to start pushing their GPU's as plug-in math-coprocessors, they'll probably start designing them to cater for that use too.
Re: (Score:2)
This would be much faster than a general purpose CPU for all sorts of machine learning concepts like hidden markovs, computer vision, speech recognition... if only I had one of the cards...
Re:what does it DO? (Score:4, Informative)
Do you want to use this to handle AI?
It depends on what kind of AI you are talking about. Path finding actually maps nicely to the GPU. AMD released a demo that showcases this by running a path finding simulation on the GPU for several tens of thousands of agents. Read all about it in Chapter 3 the Advanced Real-Time Rendering course notes [amd.com] from SIGGRAPH 2008. Demo and screen shots here: Froblins Demo [amd.com]
Re: (Score:1)
Re: (Score:2)
Actually it could handle AI. If your AI project involve neural networks this could make then run very fast. Also it could do some rather simple everyday tasks well too such as transcoding media files, adjusting the color of images and the first levels of processing for voice recognition. All of these tasks involve massive numbers of simple calculatons
Re: (Score:1)
An open spec for crunching a shitload of calculations on a GPU on a card that handless dedicated rastering graphics...
Call it far-fetched but... do I smell the opportunity of real-time ray tracing games running on top of rasterised 3D desktop?
math co-proc have always been there! (Score:4, Informative)
they just have been integrated into the main chip
by 486 era if I remember correctly.
By that time they had enough transistor to just put everything inside the same silicon chip, faster, cheaper.
Today, every CPU have an IEEE floating point unit.
To say we don't have maths co-proc is misleading.
Re: (Score:2)
And eventually what happened with the FPU will happen with the GPU.
Re: (Score:2)
intel had 486's without math coprocessors integrated (the 486sx), though it was actually on the die but disabled (either intentionally or from chips with defective math coprocs). They had a 487 you could couple with your 486sx that was still basically the same chip as the 486dx, and I think may have just disabled the other 486sx and run everything on that one processor. From what I've heard, all that craziness was due to marketing rather than technical reasons.
Re:what does it DO? (Score:4, Insightful)
The math co-processor wasn't made obsolete. It became so vital to system performance that Intel and friends started including it in on the CPU proper. These days, they call it an FPU.
Re: (Score:2)
A more cynical view is that as they became more popular, Intel started losing marketshare to alternative math coprocessor vendors, which they ended by putting the co-processor onto the CPU, which made it much harder for alternatives to compete.
Re: (Score:2)
Um, the math co-processor never became obsolete; they started building it on-die rather than in a separate package.
And yes, it is similar to CUDA and CAL, but designed for any general purpose parallel computing, not just GPU from what I can tell.
Re:what does it DO? (Score:5, Informative)
is this simply a spec that people expect ati and nvidia to conform to? or is this another api outside of CUDA and CAL, that wraps the two up so that a single api can execute code on all GPGPU's?
It's the latter: a single API + kernel language for any GPU. Because both NVIDIA and AMD are represented in the contributor list, it actually has a chance of being adopted.
Re:what does it DO? (Score:5, Informative)
It's the latter: a single API + kernel language for any GPU. Because both NVIDIA and AMD are represented in the contributor list, it actually has a chance of being adopted.
According to heise.de [heise.de] (in German), nVidia says that OpenCL applications will run seamlessly on any gpus with a CUDA-compliant driver. Does anyone know if that applies to the proprietary Linux drivers?
If this really takes off, how long until the hardworking people from the x.264 or VLC or ffmpeg or mplayer projects can write a H.264/AVC decoder that uses the GPU?
Re: (Score:3, Informative)
Yes, there is a CUDA driver and SDK for Linux on NVIDIA's site: http://www.nvidia.com/object/cuda_get.html
Re: (Score:3, Insightful)
Re: (Score:2)
Proprietary Linux drivers do CUDA, don't they? If I'm not wrong, NVidia proprietary Linux drivers do not lack of any features .... why should they lack now?
Re: (Score:1)
Re: (Score:2)
Who cares about the decoder? 2.0 GHz Core 2 Duo can handle the decoding without breaking a sweat.
That's nice and well when we're talking about desktop systems. But think about MythTV media center PCs - if you could combine an Atom CPU and a passively cooled nVidia or AMD GPU, a super-silent, HDTV capable home-grown set-top box would be possible.
Of course, an OpenCL encoder would help, too, for this kind of setup - broadcast TV encoding, for example.
Re: (Score:2, Informative)
This isn't necessary for that because modern GPUs already have dedicated hardware for video decode [slashdot.org].
Re: (Score:1)
That's an nVidia only API...
Re: (Score:2)
And appears to be profile dependent - my videos don't match their canned profile nor do I see myself changing. My fingers are crossed they can expand that....
Re: (Score:1)
Re: (Score:2)
But if you intend to play games, VIA Nano with a possible GeForce chipset (sometime soon!) would be way better (or at least the CN896 chipset and a low-power GeForce).
ARM is low power; but other than that, what advantages does it have? Does it really scale to desktop performance? I mean, I'd love to have a small board like the BeagleBoard, but the thing isn't that great for desktop performance. Where are ARM's upsides?
Pretty ignorant statement based on what I've seen (Score:2)
Oddly the XBMC folks have found that a 3ghz DUAL CORE CPU - Intel - is what it takes to do high bitrate HD video. So I'm sorry to burst your bubble but some GPU help would certainly be helpful! I'll agree that encoding takes far longer but playback shouldn't require such beefy hardware either and encoding need only be done once as opposed to DEcoding which happens on every user's machine who wishes to see the content. Have you looked at the sorts of speed-ups the NVIDA code has been giving? When you see a m
Re: (Score:2)
http://www.badaboomit.com/?q=node/4 [badaboomit.com]
Hardworking people at VLC you say? Bah. (Score:1)
VLC is fine if you don't care about preserving the quality of the format, or if you're too braindead to install proper codecs. Or if you want your integrated subtitles to look like shit, unless you run a nightly build with a few tweaks.
Honestly, I've found that Zoom Player's codec downloader and auto-configured silent install work the best for everyone, from the common person to the hardcore encoder to the obscure format enthusiast. Its a nice little stand-alone exe that, when run, will actually update your
Re: (Score:2)
Just out of curiosity, what about ATI?
(9600GT still looks like a good deal, but the HD4670 might just be slightly better for me at the moment... But do these include something like PhysX? Is it even a consideration?)
Re: (Score:2)
I'm more interested in when Matlab or the rendering-engines for applications like Maya and Pro-Engineer will add support for OpenCL. =)
Re: (Score:2)
Looks pretty much empty....
Re: (Score:1)
Microsoft isn't one of the companies listed. Which is probably a good thing, since they have a tendency to prevent progress then release their own proprietary version (ala OpenGL).
I hope this becomes a cross-platform thing. (Score:4, Insightful)
There's no way I'm writing a single line of CUDA code when it only works with nVidia hardware, and I think there are a lot of other people like me. This could open up GPGPU programming to a much wider group of programmers.
Re: (Score:3, Insightful)
CUDA on ATI would help me a bit. But a quick googling of this concept turned up a bunch of pages saying it'll never happen, doesn't work, etc. Could you post a link?
No way for CUDA on ATI (Score:3, Interesting)
CUDA on ATI can't be done easily.
As a writer of CUDA code, I can tell you that a lot of CUDA isn't as high level as nVidia's marketing would like you to believe. And thus is very much linked to specific properties of the current hardware from nVidia.
The lower-level of CUDA enables the programmer to do some really clever optimisations. But as the hardware peculiarities aren't abstracted away, writing a compatible implementation for chips from a different manufacturer which aren't exactly the same underneath
OpenCL : looks actually promising. (Score:2)
Sorry for the self reply.
Because it is supposed to be vendor neutral, OpenCL looks promising too, but I haven't read the specifications yet.
I did read the OpenCL specifications afterwards, yesterday night. Call me crazy, but that's the field I'm currently working in.
OpenCL *is* very promising, indeed.
It's not quite that different than CUDA.
What about other chipmakers? (Score:3, Interesting)
While I see quite a few members that I wasn't expecting (Creative Labs), my concern is that there are some companies that should definitely be participating in this but aren't.
By that I mean gfx chip makers such as Via or S3, as for now it seems we're tied to the major players (nVidia, AMD, Intel) for desktop/laptop implementations and that's never good for the consumer.
Either way the spec itself is a great initiative and I can't wait to get my hands on beta bulids of Snow Leopard to try it out...
Re: (Score:3, Interesting)
Oops.. I just noticed S3 is on the list, they managed to get a lot of companies on board after all.
Re: (Score:2)
Well, it doesn't mean a thing. You know, M$ was on ODF OASIS board quite some time too...
To me litmus test of OpenCL would be independent (from video card vendor) portable implementation which runs on Linux.
Participation doesn't mean a lot to how the spec would develop.
To put it in another words: I'm waiting for reaction from Mesa and X.org folks. Then, if reaction would be positive, the news would get me excited.
Re: (Score:3, Interesting)
Open Source means all of us. X.org and Mesa don't have a magic cow that shits code, it has to be written by people in their spare time (and X.org is stagnating due to a lack of developer interest). Nobody on the OpenCL list particularly give a shit about linux, and adoption will happen with or without linux or open source. Instead of waiting for other people to tell you how to feel, maybe you should sit down and read the spec.
Re: (Score:2)
That explains sooooo much.
Re: (Score:2)
I don't think your litmus test is reasonable. Because of the "big 3" video vendors signed on, virtually all modern Linux desktops will end up with OpenCL supported, and in the case of Intel and perhaps AMD, with open source drivers. nVidia will hold out open sourcing, but not many people will care because the performance will be good anyway.
Having an independent implementation here is not important, and not at all useful, given that you'd be independent for the sake of it, rather than letting the vendor wri
Re: (Score:2)
What does this have to do with the Mesa and X.org teams? They're graphics guys, and this isn't for graphics. I'd like to see inclusion in GCC (auto-vectorization maybe?) though.
Re: (Score:2)
Access to resources of video card has to be synchronized.
Unless of course you are going to "init 1" just to run you suppa-puppa OpenCL application.
You can't avoid X.org - because they are the driver of video card.
You will stuck into Mesa folks - because they are closer than anybody else to what Kronos groups is doing.
Re: (Score:2)
Does Via make anything with enough power for this to matter? All their graphics (and for that matter, everything else) seems to be bare minimum or less hardware that competes on price and power.
Don't get me wrong, I love them* but I have a hard time imagining it'd be worth anything to exploit the power in their graphics.
*Except that they don't make a server quality chipset to go with their processors. That pisses me off.
Re: (Score:2)
Does Via make anything with enough power for this to matter? All their graphics (and for that matter, everything else) seems to be bare minimum or less hardware that competes on price and power.
Don't get me wrong, I love them* but I have a hard time imagining it'd be worth anything to exploit the power in their graphics.
*Except that they don't make a server quality chipset to go with their processors. That pisses me off.
They own S3.
Re: (Score:2)
Isn't it an open standard?
Then those companies can implement support for their chips too...
They just can't influence how the standard looks or works.
Great! (Score:4, Insightful)
Now, if only they could do the same for OpenGL... Which is needed by a lot more people, and is in my opinion a lot more important for anyone who wishes to be free of Windows.
Re: (Score:1)
try this in command:
start c:\
start "C:\program files"
???
start /?
and laugh...
Re: (Score:1, Informative)
Sorry, your joke is that the utility functions exactly as documented? Really?
Re: (Score:2)
Re: (Score:2, Insightful)
...to be free of Windows
Sounds like windows is a disease. Come on don't be so critical
Windows is a condition, rather than a disease. It has unpleasant and often expensive consequences (spyware, antivirus subscriptions, etc.) for those afflicted with it and for many others (spam botnets, net worms, etc.). Luckily, it is avoidable and curable in many (but far from all) cases: just use BSD or Linux.
Re: (Score:2)
Now, if only they could do the same for OpenGL.
Care to elaborate?
OpenGL is quite well supported on both Mac OS X and Linux. So with OpenGL you are already free.
Re:Great! (Score:5, Insightful)
He wasn't implying anything about windows + GL as such, more making the observation that openGL is vital to Mac/linux - and as such those OS's are very much at the mercy of the Khronos group's actions (or more accurately - no action at all as was the case with GL3).
Re: (Score:2)
I think the OP meant, "If they could finally get around to ratifying an openGL 3.1 specification in 6 months (instead of being 2 or 3 years late as GL3.0 was); turn it into a useful standard that people actually want to use (which GL3.0 is not); and finally make good on all the things we were promised for 3.0, which they ended up ditching at the last minute. If that happens linux/mac openGL developers around the world will feel less dirty than they do right now"....
He wasn't implying anything about windows + GL as such, more making the observation that openGL is vital to Mac/linux - and as such those OS's are very much at the mercy of the Khronos group's actions (or more accurately - no action at all as was the case with GL3).
Thank you sir. That is indeed exactly what I meant, except that you phrased my frustrations a lot better than I did.
Re: (Score:2)
I think the OP meant, "If they could finally get around to ratifying an openGL 3.1 specification in 6 months (instead of being 2 or 3 years late as GL3.0 was); turn it into a useful standard that people actually want to use (which GL3.0 is not); and finally make good on all the things we were promised for 3.0, which they ended up ditching at the last minute. If that happens linux/mac openGL developers around the world will feel less dirty than they do right now"....
He wasn't implying anything about windows + GL as such, more making the observation that openGL is vital to Mac/linux - and as such those OS's are very much at the mercy of the Khronos group's actions (or more accurately - no action at all as was the case with GL3).
The 6 months for OpenCL is based upon 3 or more years of Apple work flushed out and the 6 months are the flushed out specs that satisfy all the big 3 GPU vendors and the more specialized vendors. Perhaps now that this is done, Apple will do more of the heavy lifting and then accelerate OpenGL 3.1 to be done soon as well, seeing as how much it will play a role in Snow Leopard with OpenCL.
Re: (Score:1)
The OpenCL spec is also way shorter than the OpenGL spec, and so are the header files. If you want to compare the time it takes to bring specifications out you have to compare the resulting size of these specifications.
Re: (Score:2)
Actually, that's kinda the point. OpenGL needed to go much more general purpose (read, push a load of verticies and then do general purpose computing on them to turn them into colours). OpenCL is rather more flexible than CUDA in that it can read from vertex buffer objects, and write to framebuffers, which means that it could sensibly be used for "software" graphics engines.
Re: (Score:2)
You've got that exactly backwards.
It's OpenCL that's trying to follow in the footsteps of OpenGL, not vice versa.
OpenGL is an open specification that has many implementations taking rather good advantage of graphics hardware! ;-)
OpenCL wants to establish a similar standard but for GPU based kernel execution rather than graphics rendering.
Namespace conflict (Score:1)
Re: (Score:2)
Re: (Score:2, Insightful)
OpenCL code is a subset of c, with an API that the GPU must implement. So it will work on Cell (or any CPU) with gcc and a library implementing the API.
Re: (Score:2)
You think we could get this little bugger working on Linux (and BSDs, etc) then?
http://www.leadtek.com.tw/eng/tv_tuner/specification.asp?pronameid=447&lineid=6&act=2 [leadtek.com.tw]
It looks very nice, being a 4-core Cell CPU on a card with 128MB of XDR. PCIEx1 (it's too bad I can't find PCIE splitters/risers...) and half-height but it uses a fan (I'm sure we can fix that. heh.) so it's almost the right card for a mythTV backend (or combo, or frontend since it can decode video).
It would also be nifty for those of us
Re: (Score:2)
The Cell architecture is mentioned in the intro section of the spec as an example of what OpenCL might be good for. The language is generic enough that most any vector processor should work as a target.
Moore's Supercomputer. (Score:2)
Actually I'm glad to hear this. With IGP and Crossfire/SLI with dual-GPU in a quad arrangement. One could have an inexpensive (relatively speaking) supercomputer under their desks. Throw in the upcoming quad-core in 45nm fabrication and now is a good time to be into computers.
Re: (Score:2)
Khronos isn't in the business of implementing their specs. They just define them. And OpenVG was just updated.
That's nice (Score:1)
Now, work on getting double-precision.
Re: (Score:1, Informative)
It's in the spec as an extension. The feature will appear when hardware implements it.
Plugins (Score:1)
Plugins for ffmpeg or any other codec for that matter could provide a huge boosts for linux. What about blender or yafray support.
There would be a huge uptake of Linux for video editor and 3d graphics artists.