


Transcoding in 1/5 the Time with Help from the GPU 221
mikemuch writes "ExtremeTech's Jason Cross got a lead about a technology ATI is developing called Avivo Transcode that will use ATI graphics cards to cut down the time it takes to transcode video by a factor of five. It's part of the general-purpose computation on GPU movement. The Aviva Transcode software can only work with ATI's latest 1000-series GPUs, and the company is working on profiles that will allow, for example, transcoding DVDs for Sony's PSP."
This would be great for MythTV.. Linux support?? (Score:5, Insightful)
My educated guess is, No, there won't be Linux support..
ATI was the leader in MPEG2 acceleration, enabling iDCT+MC offload to their video processor almost 10 years ago. How'd that go in terms of Linux support, you ask? Well, we're still waiting for that to be enabled in Linux.
Nvidia and S3/VIA/Unichrome have drivers that support XvMC, but ATI is notably absent from the game they created. So, I won't hold my breath on Linux support for this very cool feature.
Re:This would be great for MythTV.. Linux support? (Score:2)
Re:This would be great for MythTV.. Linux support? (Score:2)
Re:This would be great for MythTV.. Linux support? (Score:3, Informative)
Re:This would be great for MythTV.. Linux support? (Score:5, Interesting)
Re:This would be great for MythTV.. Linux support? (Score:2)
Re:This would be great for MythTV.. Linux support? (Score:3, Informative)
Pick up a 5200FX card (for SVIDEO/DVI output) and then use the GPU to do audio and video transcode. I have been thinking about audio (MP3) transcode as a first "trial" application.
"Heftier" GPUs may be used to assist in video transcode -- but it strikes me that the choice of stream programming system is most important (to allow code to move to other GP
Re: (Score:2)
Re:This would be great for MythTV.. Linux support? (Score:2)
Re:This would be great for MythTV.. Linux support? (Score:5, Interesting)
i would actually be shocked if there weren't linux support. the ability to do what they want only need to be in the drivers. i've been doing a gpgpu feasability study as an internship and did an mpi video compressor (based on ffmpeg) in school. using a gpu for compression/transcoding is a project i was thinking of starting once i finally had some free time since it seems built for it. something like 24 instances running at once at a ridiculous amount of flops (puts a lot of cpus to shame, actually). if you have a simd project with 4D or under vectors, this is the way to go.
like i said, it really depends on the drivers. as long as they support some of the latest opengl extensions, you're good to go. languages like Cg [nvidia.com] and BrookGPU [stanford.edu], as well as other shader languages, are cross-platform. they can also be used with directx, but fuck that. i prefer Cg, but ymmv. actually, the project might not be that hard, just needs enought people porting the algorithms to something like Cg.
that said, don't expect this to be great unless your video card is pci-express. the agp bus is heavily asymmetric towards data going out to the gpu. as more people start getting the fatter, more symmetric pipes of pci-e, look for more gpgpu projects to take off.
Slashdotted! (Score:4, Funny)
Re:Slashdotted! (Score:2)
What I want to see. (Score:5, Interesting)
Re:What I want to see. (Score:2)
-Jesse
Re:What I want to see. (Score:2)
One such company is Cyclone Microsystems. They offer i960 coprocessor based systems.
I don't remember the other vendor I looked at but they offered a xylinx FPGA solution or a TI DSP solution.
-nB
Already available.. (Score:3, Insightful)
Re:Already available.. (Score:3, Interesting)
Your clock advantage is about 10x [say] that is typical 400Mhz PPC vs. 40Mhz FPGA
Re:Already available.. (Score:3, Informative)
Re:Already available.. (Score:2)
The original post never mentioned ASICs that I saw. But in any case ASIC vs FPGA isn't all that relevant to the article, whereas FPGA vs Generic CPU is very relevant (and isn't at all odd). As the post you replied to said, if you do the math and it appears you can offload an operation you would normally do on your general purpose CPU to an FPGA and get the results back sooner than you could have calculated it on the CPU, it's a win (hell even if the net times are identical, you've freed up some general pu
Re:Already available.. (Score:3, Informative)
Ummm... Okay, here is a quote from the original post again... by LWATCDR:
And then a quote from tomstdenis:
tom then goes on to talk about PPC versus FPGA's, as if LWATCDR weren't talking about ASICs. Since this conversation now ivolv
Re:Already available.. (Score:2)
(order of magnitude comparison only):
Athlon 64 4000 (from pricewatch): $330
Xilinx 2.4Million gate design (from digikey): $2100-$5000.
The computing world would look a lot different if there were good $100 high-speed, high-capacity FPGAs. Now, I wouldn't argue with a good ASIC or highspeed DSP implementation for some algorithms...
Re:Already available.. (Score:2)
2.4Mgates can be quite a bit [depending on the measure of "gates"]. The typical AES-CCM core in Virtex gates is ~30k or so.
Tom
Re:Already available.. (Score:3, Informative)
You haven't specified which FPGAs you're talking about, but at those prices, you should be getting more like 6 million gates or so (e.g. an XC2V6000 goes for about $4000). Perhaps you're looking at something like a Virtex-4 FX? If so, you should be aware that what you're looking at not only includes FPGA gates, but al
Re:Already available.. (Score:3, Insightful)
Which is interesting because I'm the author of some widely deploy cryptographic software, I worked at a IP design company [for cryptographic cores]. I'd say I'm no longer an amateur when I make enough money to live on my own.
Apparently, according to you, FPGAs aren't made from silicon, they're made from fluffy bunny pixie dust
There is a strong price difference from PCB design and tapeout. I
Re:Already available.. (Score:2)
First off
Second
By your logic a bridge engineer must develop new design concepts before he's an engineer. Otherwise he's a lowly construction grunt. I'd like to think the person who designs a bridge to withstand nature, loads, etc is more than a tool swinging grunt
Re:Already available.. (Score:2)
It's sad that you feel the need to impose like you did. I guess in your important and busy life you felt it so important to stop what you're doing to try and rip on other people. I guess this is part of your "professional life" eh?
Tom
Re:Already available.. (Score:2)
I'll sulk with the knowledge mine depends on me actually being good at what I do.
I mean seriously. You shot down everything I said without offering an ounce of your credentials. That speaks volumes to your usefulness and ability.
Tom
Re:Already available.. (Score:2)
Tom
Re:Already available.. (Score:2)
You said I don't design algorithms. You're wrong I do.
You said I have no experience working with FPGAs. You're wrong. That's what I've been doing for the last 8 months.
My software is used throughout industry in products you can readily buy at any electronics store as well as inhouse business to business tools from comp
Re:Already available.. (Score:2)
No, they rule. I saw them working in heavy rain to get my feeder back on. It came back later that night. They could have easily postponed the job until the next day, but they did it.
They have a lot of work on their plate; relax, they'll get to you.
-Z
Re:Already available.. (Score:2)
No, no, no, they rock they fixed my situation.
Charles L. Stevenson would be proud.
Re:Already available.. (Score:2)
The people in charge suck! Most homes and even mobile homes came through the storm just fine. My phone line never went down. 98% of my county lost power. The people in charge SUCK!
Re:What I want to see. (Score:2)
Seeing someone make some money off of it would be.
[fn1] - Bug in the HTML Format posting ablility-
Re: probably slow. (Score:2)
So I think that this would only work if a general purpose CPU (or GPU, for that matter) has a serious architectural wea
Re:What I want to see. (Score:2)
1) What does the API for this look like from the application perspective?
2) Top of the line (read: pricey) FPGAs are mostly in the 500mhz range right now, which is in the same range as a GPU. So unless a GPU doesn't solve the problem, why would you need this? GPUs have a design that solves #1.
FPGA's cheap. Synthesis EXPENSIVE. (Score:3, Informative)
- Tools: FPGA tools are getting better, but still suck compared to modern IDEs and software development. This might be me being jaded (VHDL can get nasty), some things like System C and others are in the infancy stage, but long ways to go here.
- Synthesis time: It can take DAYS on a very fast machine to run the synthesis that
Re:What I want to see. (Score:2)
http://www.tarari.com/products-cp.html [tarari.com]
They were in Startup Alley a few years ago at N+I demoing their cards doing Perl regex's and spam checking.
Re:What I want to see. (Score:2, Informative)
Hitachi SH4 [PS2] has 128-bit doubles. (Score:2)
The Hitachi SH4 that powers the PlayStation2 can perform 128-bit double calculations in hardware [or so I'm told].
By contrast, Sun's SPARC has a "quad" precision [128-bit] double, but it's a software implementation.
I believe the chipset that powers IBM's Z390 mainframe can also do 128-bit doubles in hardware.
Re:Hitachi SH4 [PS2] has 128-bit doubles. (Score:2)
I'm rarely impressed... (Score:3, Insightful)
Re:I'm rarely impressed... (Score:5, Interesting)
Re:I'm rarely impressed... (Score:4, Informative)
Using their own codecs (Score:5, Insightful)
I'd actually be willing to spend more than $50 on a video card if more multimedia apps took advantage of the GPU's capabilities.
You're right, I stand corrected (Score:2)
Re:Using their own codecs (Score:2)
Uh no. CODEC stands for COmpressor DECompressor. In other words, it is the CODEC that does the work, whether it's part of the encoder, the program with which the user interfaces, or a plug-in. CODECs encode/decode video formats.
I usually try not to respond to ACs (why reward cowardice?) but I don't want someone to believe your misinformation.
Re:I'm rarely impressed... (Score:2)
Re:I'm rarely impressed... (Score:2)
A) Then improve it! Quality is everything. B) Of course it matters! unless we're talking about sending someone a video email or something, and then we're probably not talking MPEG[24] but more like H.263. Granted you do make that point, but since the target is much lower-resolution this is actually less important in that area. This is more significant when people are transcoding their MiniDV-source video to MPEG2 so they can put their home m
Re:I'm rarely impressed... (Score:2)
But is it worth it? (Score:3, Interesting)
with nVIDIA's 512mb implementation of the G70 core touted to be at 550mhz core, it should theoretically thrash the living daylights out of the X1800XT.
http://theinquirer.net/?article=27400 [theinquirer.net]
the decision is between aVIVO's encode and transcode abilities for h.264, or superior performance by nVIDIAs offering?
Re:But is it worth it? (Score:3, Insightful)
I don't play the sort of games that need a graphics card over $200 to look good. I never even considered looking at the high end. However, this video encoding improvement will certainly make me do a double take. I was proud of my little CPU overclock that improves my encoding rate by 20%. But the article talks about improvements of over 500%!
Keep in mind (Score:3, Insightful)
While a select few individuals still always buy the latest and the greatest, the majority of buyers look at vid
Re:But is it worth it? (Score:3, Insightful)
Re:But is it worth it? (Score:2)
Re:But is it worth it? (Score:2)
Crippled? (Score:5, Funny)
Re:Crippled? (Score:2)
GPU or CPU? (Score:3, Interesting)
Video cards with GPU's used to be a "cheap" way to increase the graphic processing power of your computer by adding a chip who's sole purpose was to process graphics (and geometry, with the advent of 3d-acellerators).
Now that GPU's are becomming more and more programmable, and more and more general~purpose, what, really, is the difference between a GPU and a standard CPU? What is the benefit to having a 3d~acellerator over having a dual~CPU system with one CPU dedicated to graphic processing?
Re:GPU or CPU? (Score:3, Insightful)
In a few years, there will be no real benefit to the GPU. Not too many people write optimized assembly level graphics code anymore, but it can be quite fast. Recall that Quake ran on a Pentium 90MHz with software rendering. It's only getting better since then. A second core that most apps don't know how to take advanta
Re:GPU or CPU? (Score:2, Interesting)
Re:GPU or CPU? (Score:2)
Agreed. The current hardware limitation is power dissipation for both CPU and GPU. Hence multi-core from AMD and Intel.
Rendering a 1600x1200 4X AA scene with full filtering on a top tier dual core system would yield perhaps 1fps with an optimized software path.
Speculation. Besides, not many people run at that resolution with FSAA except GPU fanboys.
"You'll end up wanting a MMU" Nonsense.
I've al
Re:GPU or CPU? (Score:2)
Re:GPU or CPU? (Score:4, Insightful)
Most of that stuff can be done with OpenGL/DirectX or ray tracing. Grasses are sometimes done in OpenGL with instancing small clumps. In RT you'd use proceedural geometry or instancing.
For the snow, both renderes would probably do similar techniques.
Sand dunes - either method needs an engine with deformable geometry - both can support that.
Water simulation is something I don't know much about. For the FFT methods of simulating waves it's possible that a GPU has an advantage. Once it start interacting with objects, I don't know how people handle that.
Your quesitons all point toward vast detailed worlds with lots of polygons. RT scales better with scene complexity. To get more traditional methods to work well, you get into fancy culling techniques (HZB comes to mind) and RT starts to look simpler - because it is.
Re:GPU or CPU? (Score:5, Insightful)
1. On another note, as polygon counts skyrocket they approach single pixel size
This is not happening. Not anywhere (except maybe production rendering). It is far too time-consuming, expensive, and labor-intensive to produce huge numbers of high-polygon-count models for games. Vertex pipes are currently under-utilized in most games and applications now. Efforts are underway to allow procedural geometry creation on the GPU to better fill the vertex pipe without requiring huge content creation efforts. See this paper [ati.com] for details.
2. A second core that most apps don't know how to take advantage of will make this all the more obvious.
This undercuts the argument you make in the next paragraph. Also, it's not true. Both the PS3 and XBOX 360 have multiple CPU cores. It's true that current-gen engines aren't optimized for this technology, but next-gen engines will be.
3. multicore CPUs are nearing the point where full screen, real time ray tracing will be possible. GPUs will not stand a chance.
This might be true, but so what? Ray tracing offers few advantages over the current-gen programmable pipeline. I can only think of 2 things that a ray-tracer can do that the programmable pipeline can't: multilevel reflections and refraction. BRDFs, soft shadows, self-shadowing, etc. can all be handled in the GPU these days. Now, you can get great results by coupling a ray-tracer with a global illumination system like photon mapping, but that technique is nowhere near real-time. Typical acceleration schemes for ray-tracing and photon mapping will not work well in dynamic environments, but the GPU could care less whether a polygon was somewhere else on the previous frame.
Hate to break it to you, but the GPU is here to stay. Why? GPUs are specialized for processing 4-vectors, not single floats (or doubles) like the CPU + FPU. True, there are CPU extensions for this, such as SSE and 3DNOW, but typical CPUs have a single SSE processor, compared to a current-gen GPU with 8 vertex pipes and 24 pixel pipes. Finally, do you really want to burden your extra CPU with rendering when it could be handling physics or AI?
Re:GPU or CPU? (Score:2)
This is not happening."
Because a GPU is too hard to program for recursively refining nurbs or doing subdivision surfaces on the fly.
I'll pass on #2. I'm not sure what to say - future engines using mutiple cores effectively is somewhat speculative at this point. But lets say you win, so it takes another generation to catch the GPU.
"3. multicore CPUs are nearing the point where full screen, real time ray tracing will be po
Re:GPU or CPU? (Score:2)
Re:GPU or CPU? (Score:2)
What I want to know is whether, given the new-found programmability of the GPU, more pressure will be applied for ATI and nVidia to open up the ISAs to their graphics chipsets.
Apples and pears (Score:2)
This means that the 'general purpose GPU' code, isn't really going to be general purpose, it going to be heavily vector orientated. On the other side the CPU is more general purpose, good at running many tasks and handling interrupts &co, for this reason the CPU won't replace the GPU and the GPU won't re
Re:GPU or CPU? (Score:3)
That depends on what you mean by the "one CPU dedicated to graphic processing." If you mean something on the order of a second Pentium or Athlon that's dedicated to graphics processing, the advantage is tremendous: a typical current CPU can only do a few floating point operations in parallel, where a GPU has lots of pipes to handle multiple pixels at a time (or multiple vertexes at
Yawn... (Score:2, Interesting)
An Implementation of a FIR Filter on a GPU [sunysb.edu]
Re:Yawn... (Score:3, Interesting)
http://openvidia.sourceforge.net/ [sourceforge.net]
But I'd rather have it the other way around! (Score:2, Interesting)
So, what about it ATI? Or will thi be an NVIDIA innovation?
Re:But I'd rather have it the other way around! (Score:2)
lessons of "array processors" from 1980s (Score:4, Informative)
All these companies died mainly because the commodity computer makers could pump out new generations about three times faster and eventually catch up. And the general purpose software was always easier to maintain than the special purpose software. Perhaps graphics card software will buck this trend because its a much larger market than specialty scientific computing. The NVIDAS and ATIs can ship new hardware generations as fast as the Intels and AMDs.
Re:lessons of "array processors" from 1980s (Score:2)
The improvement on general purpose CPU were mainly gained by increase of cache size, advanced pipelining and clock increase, all these factors seems to have somewhat be exploited to the max by current CPU so now Intel and AMD have to fall back on multi-core CPUs which need special purpose software to be exploited efficiently.
Still while NVidia and ATI can
Re:lessons of "array processors" from 1980s (Score:4, Interesting)
A GPU is, effectively, a very wide vector unit (1024-bits is not uncommon). What happens when CPUs all include 2048-bit general purpose vector units? What happens when they include a couple on each core in a 128-core package? Sure, a dedicated GPU will still be faster - but it won't be enough faster that people will care. For comparison, take a look at Chromium. Chromium is a software OpenGL implementation that runs on clusters. Even with relatively small clusters, it can compete fairly well with modern GPUs - now imagine what will happen when every machine has a few dozen cores in their CPU.
Done in Roxio Easy Media Creator 8 (Score:2, Informative)
Apple's core image (Score:4, Informative)
http://www.apple.com/macosx/features/coreimage/ [apple.com]
There's a CPU in my keyboard too... (Score:2, Funny)
Re:There's a CPU in my keyboard too... (Score:5, Interesting)
Re:There's a CPU in my keyboard too... (Score:2)
He wins by using the extra memory in his watch...
Linux Support (Score:3, Informative)
Hope they get the standard right this time (Score:2)
I can't use the files I recorded on anything but ATI's software and Pinnacle Videostudio (go figure, it understands the codec).
funny about memory comments (Score:3, Interesting)
"This is, after all, one of the fastest CPUs money can buy, paired with very fast RAM.
"1 GB of very low latency RAM "
After the other review [techreport.com] posted today [slashdot.org] about fast memory doing almost nothing for transcoding:
"moving to tighter memory timings or a more aggressive command rate generally didn't improve performance by more than a few percentage points, if at all, in our tests."
"Mozilla does show a difference between the settings, both on its own and when paired with Windows Media Encoder. Still, the differences in performance between 2-2-2-5 and 2.5-4-4-8 timings, and between the 1T and 2T command rates, are only a couple of percentage points."
Re:funny about memory comments (Score:2)
if u bothered to RTFA you'd realize they're bragging about system memory, not video card.
I suppose the quote "fastest CPUs money can buy, paired with very fast RAM" should have clued you in that they were talking about CPUs and system RAM, not the GPU and RAM on the video card but if you didnt bother to RTFA it only follows that you wouldnt even bother t
Makes sense (Score:2)
Apple foes this now. (Score:3, Insightful)
Re:Apple foes this now. (Score:3, Interesting)
The idea behind using your GPU in this case is even more far reaching. While using a GPU for any visual effect is fairly logical...what about SETI@Home? What about Folding? What about for runing kalc
See the difference?
Unlikely to be cross-platform (Score:2)
At the risk of becoming -1 redundant, many other posters have already pointed out that stuff like this should be done in a generic shader language so that it can be run across a gamut of GFX cards - I'm no programmer, but in my mind this would be like current CPU apps asking "do you support MMX? SSE? SSE2?" etc etc etc. Interesting projects like LibSh [libsh.org] offer to provide a platform-independent me
Sounds useful for my AGP-based system.... (Score:2)
In the meantime... (Score:3, Interesting)
Can anyone say... (Score:2)
--dave
See Foley and Van Dam, Fundamentals of Interactive Computer Graphics Addison-Wesley, 1982
Re:GPU advantages over CPU? (Score:5, Informative)
But don't take that out of context. Ask a GPU to compile the linux kernel [which is possible] and an AMD64 will spank it something nasty. *GENERAL* purpose processors are slower at these very dedicated tasks but at the same time capable of doing quite a bit with reasonable performance.
By the same token, a custom circuit can compute AES in 11 cycles [1 if pipelined] at 300Mhz which when you scale to 2.2Ghz [for your typical AMD64] amounts to ~80 cycles. AES on the AMD64 takes 260 cycles. But, ask that circuit to compute SHA-1 and it can't. Or ask it render a dialog box, etc...
Tom
Re:GPU advantages over CPU? (Score:2)
Tom
Re:GPU advantages over CPU? (Score:2)
Re:GPU advantages over CPU? (Score:2)
Re:GPU advantages over CPU? (Score:2, Informative)
Why, yes it would. GPUs fill a For one thing, about 90% of the transistors on a GPU are used for processing. About 60% on a CPU are used for processing (the rest is used for caching).
There are also many more transistors in GPUs these days than CPUs. Graphics processing is inherently parallel and streamed. That's what a GPU does very well, very fast. Grab 8 texture samples simultaneously each clock cycle, the next stage linearly blends these floating
Re:GPU advantages over CPU? (Score:2)
Re:GPU advantages over CPU? (Score:2)
Still, it suprises me that they didn't get a quicker ray tracer than that since I seem to remember reading about
Re:Will all x1000 cards do this? (Score:2, Informative)
From the article (second page):
"The application only works with X1000 series graphics cards, and it only ever will. That's the only architecture with the necessary features to do GPU-accelerated video transcoding well."