3D Raytracing Chip Shown at CeBIT 391
An anonymous reader submits "As noted at heise.de Saarland University is showing a prototype of a 3D Raytracing Card at CeBIT2005. The FPGA is clocked at 90 MHz and is 3-5 times faster in raytracing then a Pentium4 CPU with 30 times more MHz.
Besides game engines using raytracing there was a scene of a Boeing with 350 million polygons rendered in realtime."
Hardware encoding (Score:5, Interesting)
I am really not surprised at the performance as most anytime you build code into hardware, it is significantly faster. For instance, I used to have a Radius 4 DSP Photoshop accelerator card in my old 68030 based Mac IIci I bought in 1990 that would run Photoshop filters significantly faster than even my much later PowerPC based PowerMac 8500 purchased in 1996 with faster hard drives and more memory.
The same sorts of benefits can be seen in vector math for optimizations that have been built into the G4 and G5 chips with Altivec [wikipedia.org].
So, the question is: Can these guys get ATI or nVidia to buy their chip?
Re:Hardware encoding (Score:2)
Re:Hardware encoding (Score:5, Insightful)
Re:Hardware encoding (Score:2)
Re:Duh, you develop with FPGA. Then you make ASIC. (Score:3, Insightful)
(I work for an ASIC design team for a living, so yes, I do know what I'm talking about.)
Re:Hardware encoding (Score:5, Informative)
They are trying (surprise, surprise).
From their site (already melted (yes, yes, mirrordot)): We are very much interested in evaluating new ways for computer games and therefore like to cooperate with the gaming industry. Thus if you are in such a position, please send us an email!
CC.
Re:Hardware encoding (Score:4, Informative)
So to sum up, don't expect to see vastly specialised GPU's for raytraycing hitting the market, at least not for the mainstream buyer. It's more likely that we'll see GPU's become more generalised to the point where raytracing can implemented on software. Will they be as fast as a purpose built chip like this? No, more than likely they won't. Will developers be able to do a whole lot more with them? Most definitely, and though that will come at a significant performance penalty for the moment, I think it's the right trade off to make as we should see far more creative uses of hardware put into practice, such as work being done already to use GPU's for something other than Graphics Processing.
Re:Hardware encoding (Score:2)
Re:Hardware encoding (Score:5, Interesting)
Think of it like this... Unreal 4 or whatever the next next gen will be decides to partner up with these guys. They develop an engine that runs at 60fps with amazing graphics, etc. You can buy the USB3 or FW1600 or whatever add-on needed for the game for, say, $50, or a bundle for $75 that has the addon and the game. The development cycle would be much easier as there is only one type of hardware to worry about, and the consumer would win as they could get the new hottness game without having to drop $300 on a new new video card.
It could also serve as an amazingly effective copy protection scheme. Can't very well play the game without the required accelerator.
Seems possible to me.
Re:Hardware encoding (Score:2)
They were never very popular then. I can't imagine them becoming so now.
Re:Hardware encoding (Score:3, Interesting)
Re:Hardware encoding (Score:2)
I'm sure they probably still do this, but it's been some time.
Re:Hardware encoding (Score:3, Interesting)
At the end of the day, there comes a point where the software checks it's key/dongle/word from the manual/price of fishcakes in japan and asks itself "So, am I allowed to run?" and all you need to do is ensure the answer to that is always "Yes!"
Do you know what an FPGA is? (Score:5, Interesting)
This could not be further from the truth. FPGAs are more flexible than any of their counterparts. FPGA stands for "field programmable gate array," and are basically a matrix of memory elements (at the very least latches) connected to gates that configured to be a particular type of gate via a ROM or something similar.
It's kind of like a chip emulator written in hardware. You may be wondering why we don't use these all the time. First, they're a lot more expensive, bigger, and more power consuming than their one-chip cousins. Second (as if that isn't really enough), they're usually 2-5 times slower than the same logic on a custom chip.
So the big question is why should we use them? What improvements can they give that normal chips can't?
The big gain is when you want to optimize the hardware for a specific application and be able to change it. These were used in high end digital video cards to be able to handle whatever kind of signal is actually output by whatever kind of camera you've got (I can only assume this is still the case, but I stopped keeping track about 2000).
I don't know if the people who wrote this thing take advantage of this idea within their design, but it's a possibility.
Re:Do you know what an FPGA is? (Score:2, Insightful)
From what I've learned in one of my digitial system design courses from college, ASIC have an upfront production cost of $1,000,000 to $2,000,000 and a small unit price, while FPGAs have a much lower fixed cost (~ $10/unit).
So, I imagine they will manufacturer ASIC chips when they get a big sponser, unless they are using the FPGAs dynamic abilities...
Re:Do you know what an FPGA is? (Score:3, Informative)
Fortunately, there are some [actel.com] companies [latticesemi.com] that are incorporating flash memory on to their FPGAs instead of using the standard [xilinx.com] SRAM. The problem is that flash-based FPGA
Re:Hardware encoding (Score:5, Interesting)
Of course, raytracing produces beautiful results compared to the other methods of 3d graphics, but it is MUCH more expensive in terms of CPU cycles on today's CPUs and non-existant on graphics chips -- the first gfx chips were polygon-based because drawing polygons is indeed easier than raytracing even with specialized hardware. of course, specialized hardware definitely helps polygons as well. my 300mhz/whatever TNT2 can render a scene as fast as the fastest pentiums today can using software rendering.
all of the big renderfarms rely exclusively on the CPU to do their animations. this could change all that. I for one look forward to seeing the potential this has.
Re:Hardware encoding (Score:5, Insightful)
This may not be true for very long. The complexity of a scene in a traditional polygon renderer like nvidia's chips scales fairly linearly with the number of polygons in view. Not so with raytracers. They have hierarchical structures to test for which group of triagles a ray may intersect and scales more like O(n log n). They also render _only_ viewable pixels, while overdraw is a major hurdle for traditional 3d cards.
What this means is that as scenes get increasingly more complex, there is a crossover point where ray-tracing will overtake traditional rendering, and dedicated raytracing hardware ensures that this happens sooner. If you add this to the fact that raytracing lets you have perfectly smooth non-polygonized objects, perfect reflections and other features not easilly replicated by traditional rendering you'll see the incentive.
A case in point: prman, the renderer used by Pixar is a traditional polygon rasterizer, but Pixar has on occation used BMRT (A renderman compliant ray-tracer) for scenes that require ray-tracing for realism.
Specifically a scene in A bug's life depicting a glass bottle filled with nuts was renedered using BMRT. Flexible and robust realistic reflection and refraction is solely in the domain of ray tracing. What you saw in Half Life was only cheap tricks which would fail miserably in less constrained scenes.
Re:Hardware encoding (Score:5, Informative)
I assume you're talking about kd-trees... these do indeed offer very nice performance characteristics, but they're designed for static geometry. Efficient raytracing for dynamic geometry (moving or deforming objects) is AFAIK still far from "solved".
If you add this to the fact that raytracing lets you have perfectly smooth non-polygonized objects
and take away the fact that they don't particularly like the arbitrary triangle meshes that make up the vast majority of real datasets...
Flexible and robust realistic reflection and refraction
Yes, "Flexible and robust" is the killer. And not just for refraction/reflection; there's still no fully-general, clean, robust method of shadowing for rasterizers, and it's not for want of trying. Radiosity is a joke. Attempts to get realism out of current rasterization approaches are bodges piled on kludges piled on hacks. It became clear some time ago that the technology was heading up a dead end. Of course, so much has been invested in making that dead-end fast that it's going to be hard to take the performance hit of moving to a better but less optimized approach.
I suspect we'll eventually end up with a hybrid, rather like current deferred-shading techniques. It'll be interesting to watch it all pan out.
Re:Hardware encoding (Score:3, Interesting)
while overdraw is a major hurdle for traditional 3d cards.
Not lately (as of geForce5K+). There are two very tried-and-true methods of getting around this. One is to lay down a depth-only pass first, which allows the hardware to render at ~2x the normal polygon throughput (which is pretty much infinite at this point anyways), and then reject the hidden color pixels *before shading ever occurs*. Also, there is *a s
Re:Hardware encoding (Score:3, Interesting)
Are not just polygon pushers since a long time now. You should watch Doom3, maybe.
For example, with nvidia shaders 2.0, in a nvidia 5900, there's some people who built a photon mapping kind-of-realtime renderer. I'm sure with the new cards, it could be actually realtime.
For your information, ray tracing is far, far from providing the best results, part
Re:Hardware encoding (Score:2, Insightful)
Whoa, 90 MHz?? (Score:3, Funny)
Not fast for a computer... but for a chip... (Score:2)
I can't comment too much since we're suffering from the slashdot effect (can RTFA but the pics aren't loading), but depending on chip size one could imagine this being very useful on a handheld device or something similar. Perhaps a small integrated graphics ta
meh (Score:4, Funny)
The FPGA is clocked at 666 MHz and is 3.5 billion times faster in raytracing then a 80486 CPU.
Besides game engines using raytracing there was a scene of Cthulhu awakening in 350 trillion polygons rendered in realtime.
i.e. Vaporware
Re:meh (Score:3, Funny)
Re:meh (Score:3, Funny)
Re:meh (Score:2)
I mean, what else is there?
Re:meh (Score:2)
Re:meh (Score:2, Funny)
Just annonced (Score:5, Funny)
What I would like to see (Score:5, Funny)
oh yeah (Score:3, Interesting)
it's amazing to me that nvidia have ignored this up until now, their existing simd architecture and memory subsystems can be easily adapted...
all we need now is consumer push!
How, Exactly? (Score:2, Interesting)
Only thing I can think of is that ray tracing would
allow us to replace complicated hacks for shadows
and reflections with a more natural implementation, but I can't imagine how this will usher in a new era of gaming.
Re:How, Exactly? (Score:3, Interesting)
Or how about global illumination lighting effects? Truely emissive surfaces and area lights? As a hobbyist map maker, I would kill to have an engine that supported these; imagine being able to just tag the sky as an emmisive surface and have the entire level li
Re:How, Exactly? (Score:3, Interesting)
You have to use a different kind of rendering to get soft shadows and proer reflection. Radiosity is a popular method. Basically you treat all objects as light sources, after a fashion. L
Re:oh yeah (Score:5, Interesting)
By making a card that rendered in immediate mode, nVidia had, ya know, a market. If they created a raytracing card, they would have needed to invent a new API to run it. They would have been the only ones with a card that used the API. Because they would have had a very small installed base, nobody would have written programs to take advantage of the API. Other companies have made raytracing accelerators. This isn't new. Most of them have not done incredibly well because there is so little actual use for the product.
Think of it this way... How many programs have you seen written for the 3DFX glide API? So, if you are one of the people who still has a glide card, but it was designed so that it couldn't do OpenGL becuase it used completely different technology, how useful would it be to you?
Personally, I'd love a card like that, if it was well supported by Lightwave, and had a vibrant developer community, and multiple vendors making cards for the raytracing API, and I was sure it wouldn't disappear soon.
Re:oh yeah (Score:2)
Normally I avoid Grammar Nazi posts on slash, but this is completely indecipherable. If you're setting up a hypothetical, use the subjunctive. If you're asserting that Glide-oriented cards don't support OpenGL, you're flat-out wrong; they have full OpenGL support. 3dfx also provided a stripped-down, "Quake only" driver
Re:oh yeah (Score:2)
Re:oh yeah (Score:3, Interesting)
Raytracers can easily do hard shadows, reflection, refraction, and order-independent transparency. Today's rasterizers can do almost all that too: hard shadows (stencil shadows), and "good enough" reflections and refractions (using environment maps and shaders). Order-independent transparency is a tough one; it can be kludged using shade
Impossible! (Score:5, Funny)
That's ridiculous. Everyone on slashdot surely knows by now that the only reliable way to compare processor speeds across architecures is to compare clock speed!
Re:Impossible! (Score:2, Funny)
--
Free iPod? Try a free Mac Mini [freeminimacs.com]
Or a free Nintendo DS, GC, PS2, Xbox [freegamingsystems.com]
Wired article as proof [wired.com]
Re:Impossible! (Score:2)
Wait, what?
Re:Impossible! (Score:2)
Can someone setup a torrent (Score:2, Informative)
Re:Can someone setup a torrent (Score:2, Interesting)
Torrent of low quality up, others will come as they finish downloading.
Open FPGA? (Score:3, Insightful)
Re:Open FPGA? (Score:3, Insightful)
This is a major problem. A modern processor can abstract away many of the complexities of its design. The ISA just has to run, and the processor can handle the details of how this happens. With an FPGA, you are down-n-dirty with the hardware. Any architectural changes have an major impact on how you handle t
Sweet deal! (Score:2, Informative)
Re:Sweet deal! (Score:2)
Should run Doom3 pretty well then I suppose...
Re:Sweet deal! (Score:5, Interesting)
In the simplest algorithm, assume only point lights, no spots or area lights. Basically, when you are shading a point, so you can draw it on screen, you trace a ray from that point to each light. (You may limit the lights that are at a distance beyond some cutoff, doesn't matter.) If the ray hits some geometry on the way to the light, it is in shadow for that light.
So, without reflections, or anything cool, just pointlights, and shadows, you will trace
S+L*S rays
where S is the number of scene samples (pixels) that you are shading, and L is the number of lights. The lone S comes from all the rays you trace from the eye point out into the scene in order to figure out which point is visible at which pixel.
If you have lots of reflections and refractions, that's what can really start to slow things down. At your point being shaded, you have to trace a ray each for the reflection, and for the refraction. If the reflection ray then hits another surface which is reflective, you trace another ray to get the reflected reflection, same with refraction. So, in theory, each sample point can spawn two new rays in addition to the rays for shadow tests, and each of those two new rays can result in two more new rays, etc. You basically have to set some limit to how many times you let it recurse, because two parallel mirror planes would take forever to render accurately.
But wait, there's more! (it slices, it dices!) Everything really starts to explode when you throw out soft shadows and hard reflections. If you want everything to be nice and soft and smooth, you basically have to trace lots of rays and average the results. So, instead of each recursion in a shiny refractive scene spawning two more rays, it may need to spawn 20 or 200. Assume a max recursion of 5, and 20 rays being generated by each shading point.
First point traces twenty rays.
Each of those 20 trace 20 for 400.
Each of those 400 trace 20 for 8000
160,000
3,200,000 shading sample points for the fifth level of recursion, each of which will need to trace rays for each of the lights which might not be casting a shadow on it, possibly many more for soft shadows.
So, 3.2 million times Lights times soft_shadow_samples times pixels times samples_per_pixel (and believe me, 10 samples each for the reflection and refraction is not very smooth in my experience!)
A veritable explosion of rays, as I am sure you see. I won't even begin to discuss radiosity, because that's actually slow, and computationally intensive.
Now, we get to the subject of Doom3... I'm not sure this hardware would actually be that well suited to Doom3. You know all the lovely shading effects, with detailed highlights and bump mapping? They pretty much define the Doom3 Experience. That all comes from a technology called programmable shading. Basically, while your GPU is rendering the polygons in the game, it runs a tiny little program that determines the precise shade of black for each and every pixel.
A raytracing accelerator takes advantage of the fact that ray hit-testing is a very repetative chore which can be done in hardware very efficiently.
But, as you can see, most of the really interesting rays in a scene are the "secondary rays." The rays that are for reflection and refraction and lighting and such. So, suppose this card calculates a ray, and figures out the point that needs to be shaded. Because the accelerator is all in hardware, for programmable shading like Doom3, it would need to hand-off back to the host processor, which would run the shader code, which would ask for 20 more rays, etc. So, with a fully fixed function raytracer, there would either be annoying limits on what the scene could look like, or you would constantly be going back and forth be
raytracing advantages (Score:5, Insightful)
and iirc raytracing is a very simple thing to parallelize. given the performance they are getting out of their FPGA prototype, I expect this will scale nicely.
imo raytacing is the obvious future of graphics cards.
as an aside, a lot of game mechanics is dependent on raytraces for detecting collisions. now if you could use a raytracing GPU to handle that as well, you've offloaded yet more work from the CPU...
Performance (Score:5, Insightful)
On the Ray Traced Quake 3 Website it says that runs faster with more computers (about 20 fps@36 GHz in 512x512 with 4xFSAA)
Assuming that is correct,a normal chip can render Ray Traced Quake 3 like graphics at 2 to 3 fps on a 4GHZ machine which means the Ray Tracing Chip could do it at 6 to 9 fps. This might be real-time for alot of research, but when it comes to games anything less then 15 fps is a joke. I'll be interested when they can hit 30 fps, with more graphics complexity then Quake 3.
Re:Performance (Score:2, Informative)
Re:Performance (Score:3, Insightful)
Re:Performance (Score:2)
Research not meant to be immediately practical (Score:3, Insightful)
While it's certainly not enough to start playing games, it's a heck of a lot closer than I thought was possible. And there's a lot of tweaking that could be done to speed the process up with present technology. An FPGA [wikipedia.org] is the integrated circuit equivalent of a stack of lego. It makes it possible to build custom hardware without forking out for a custom chip; they are however much slower than such a custom-built. I expect if Nvidia decided to mak
Re:Performance (Score:2)
It probably wouldn't be a huge challenge for them to up their clock rate by 75% or so to get 15FPS. Imagine if they moved it from FPGA into custom silicon with appropriate chip designer voodoo.
FINALLY (Score:4, Interesting)
Every time I hear about the latest beast of a GPU from ATI or NVidia, I can't help thinking what a waste all those transistors are for anything other than gaming, and maybe a couple other applications. We should be putting those resources into an array of runtime-programmable FPGAs! Your computer could reconfigure itself at the hardware level for specific tasks -- one moment a dedicated game console, the next a scientific workstation, etc.
Lest I get too breathless here, does anyone care to inject some reality into this? Are there technological reasons why FPGAs haven't burst into the mainstream yet, or is it something else? Have I misunderstood their potential entirely?
Re:FINALLY (Score:2, Interesting)
Re:FINALLY (Score:2)
An extra reason for the slashdot crowd would be the open source potential. Wide availability would attract developers to open source FPGA projects and could lead to
Price.. and compile time (Score:5, Informative)
There's also the speed issue - I've spent DAYS of CPU time to get a design syntheized from VHDL for a moderately complicated IC built up from available cores.
Factor in optimizing floorplans and the like, and you're talking about serious time commitments to optimize the hardware.
It works; I've been paid to do it in the past; but it's not something I can see in the consumer market for the time being.
An exciting hybrid is intersting though, putting silicon CPU cores on the same die with an FPGA. They've been around for awhile, and I haven't done any FPGA projects in ~18 months - but I haven't seen any real movement outside of areas where FPGAs are already popular.
See Open Cores [opencores.org] (no, not sores..
Re:FINALLY (Score:2, Informative)
FPGAs are not a good tradeoff (Score:2)
People build special-purpose chips for the mass market because it's the cheapest way of getting the functionality out.
And as general purpose computational devices, FPGAs are just too hard to program and too inflexible.
Re:FINALLY (Score:3, Informative)
high quality animation (Score:4, Interesting)
why not old crap pcs are clusters (Score:3, Insightful)
Hell, even 20 chipped xboxs
Re:high quality animation (Score:2)
Saarland... (Score:4, Interesting)
It's rather pretty in a European countryside kind of way - hills with wine grapes on them, big rivers with boats cruising up and down, and big vegetable gardens everywhere (Germans sure love their vegetable patches) - though I doubt it's the kind of place too many international tourists visit. Not the kind of place you'd expect cutting-edge graphics research either; but then, you find all manner of interesting research in all manner of places. Even Melbourne, Australia :)
Hi to any residents of Saarland reading this - are they holding the German round of the World Rally Championship there this year?
Re:Saarland... (Score:2)
(another place where bratworst are consumed)
SFA in melbourne (Score:2)
Sure theres a handfull of gaming companies though to get in that line of work you already have to have been in a team that made a game or 2 or 3 and have a degree (how many degree people know gaming?) I know companies want quick results with no trainups, but comon, if someone already has done 3 games, then they are probably burned out tired from that type of wor
The "Q3RT" screenshots... (Score:4, Insightful)
Call me a kid, but this amazing technology appears and all i can think is how cool would it be to see enemies coming behind you reflected in a sphere...
Too bad there's no video - but then again, the poor server is doing bad enough as it is.
Re:The "Q3RT" screenshots... (Score:2)
That is pretty sad. Because I'd be thinking about shooting a laser rifle at that ref(le|ra)ction an seeing what happens!
Game rule 1, no direct shots of the enemy!
Fun fun.
Re:The "Q3RT" screenshots... (Score:2)
What kind of FPGA? (Score:3, Interesting)
Re:What kind of FPGA? (Score:3, Informative)
it's a virtex II 6000-4
from some pdf at http://www.saarcor.de/pubs.html [saarcor.de]
raytracing with 350 million polygons? (Score:3, Interesting)
Are you sure the Boeing thing was raytracing 350 million polygons? Or just traditional raster pipeline rendering?
See, the reason I ask is, you generally get away from raytracing polygons and raytrace against the actual nurbs or other mathematical surface definitions. That's the point. You don't feed it to simple scan-and-fill raster pipelines.
Re:raytracing with 350 million polygons? (Score:3, Informative)
Re:raytracing with 350 million polygons? (Score:2)
If it's true that they're only tracing triangles, then this isn't "real" raytracing to me. I'm sure it will be cool and useful, but there's a lot of things you can do with ray tracing that require more than triangles. Raytracing against solid objects with intersections and whatnot (ala Povray) can do a lot of things that this can't (like correctly render a crystal sphere with something behind it). I do hope hardware raytracing takes off though. The 3D we have to day is all an optimization hack, hence th
Re:raytracing with 350 million polygons? (Score:3, Informative)
Anti-Planet (Score:5, Interesting)
That, along with Wolf 5k [wolf5k.com] inspired me to start working with software rendering. I think ray tracing will eventually be the way real time graphics are rendered in order to keep upping the bar for realism.
Real Time Software Rendering [icarusindie.com]
I'm working on tutorials covering software rendering topics. The tutorials start by deobfuscating and fully documenting Wolf5K, cover some basic ray tracing and are now going through raster graphics since the concepts used for raster graphics apply for ray tracing as well. I'll be returning to do more advanced ray tracing stuff later. The tutorials also cover an enhanced version of Wolf5K [icarusindie.com] written in C++ that is true color and has no texture size limitations.
Re:Anti-Planet (Score:2)
Re:Anti-Planet (Score:4, Informative)
more recent amd chips have sse2, and sse3 on amd64 is just round the corner.
Surprising (Score:2)
Re:Surprising (Score:2)
To render 2048 pixels in 41milliseconds shouldnt be too hard, thats 49 pixels in 1ms. About 1/1000th of 3ghz cpu which is about 2.5million instructions, or 51000 per pixel.
Though pixar has lots of money, and instead of just buying 270 G5 servers, they could make their own FPGAs to do it.
Re:Surprising (Score:2, Informative)
Mirror to video (Score:2, Informative)
I seriously thought PS2/Emotion engine wud do this (Score:3, Insightful)
Imagine if nvidia threw in an extra 5m transistors for a raytracing option
Re:I seriously thought PS2/Emotion engine wud do t (Score:2)
PS5, Xbox Next II? (Score:3, Insightful)
Only ten years too late... (Score:2)
HQ RT Ray Tracing (Score:3, Insightful)
The screenshots on the site all show images that could easily be rendered with much greater quality and efficiancy using shadow maps or stencil shadows and manual matrix transformations with portal rendering for reflections.
When this hardware can render scenes on par with that of a professional software ray tracer in real time, then there will be some serious consumer demand.
It takes more than a chip (Score:5, Interesting)
Once this functionality was achieved, there were some contextual architectural decisions to be made about what asic would include these gates. The company decided to implement these gates on a chip that had about 16MB of ram on it and its own execution unit (vaguely like one of the subchips in IBM's upcoming cell architecture, IIUC) and then to put arrays of these independent exec chips on daughter cards.
Many of these decisions were trying to solve the specific problems of raytracing, e.g. how do we get geometry info into the chips efficiently, how can we parallelize the running of shaders so they don't bottleneck things, etc. These problems manifested themselves quite differently than they did for zbuffering hardware, and there were lots of clever-yet-brittle constructs used which could be shown to work in specific cases but which had pot-holes that were hard to predict when scaling or changing the problem/scene at hand.
Rather than selling these chips themselves, the company decided that programming them was hard enough that the company itself would package up the chips into a "rendering appliance", which was essentially a computer running linux with a few of these daughtercards in them. For a software interface to rendering packages, the company chose Renderman. The task then became to translate rendering output from disparate sources (Maya, etc) into renderman expressions, and this was devilishly hard to get right. Each rendering package had to be individually tweaked in emulation, and some companies didn't help out much with info, and even those that did weren't able to supply all the info needed in many cases... my former employer ended up chasing un-spec'd features down ratholes.
The end result was really a disaster. Nothing worked quite right, which was problematic because these chips were marketed not just as fast but as faster drop-in replacements for existing software renderers.
I find it interesting how this entire tsunami of problems snowballed from the initial foundation of how raytracing algorithms (and therefore hardware) are different from zbuffering.
Re:It takes more than a chip (Score:3, Interesting)
Programer vs. Artist screenshots. (Score:4, Insightful)
Let me ask you. Why would you expect teams of electrical engineers/mathematicians/programmers be able to produce a prettier image than a team of extremely talented artists.
The job of the artist is to make it look pretty. The job of the guys making this chip is to provide features and to make it go fast.
So it's not fast enough yet. It's a prototype. When it can render decent resolutions at decent framerates then bring in the artists and see what they can do with it.
Re:A Boeing? (Score:4, Informative)
Re:A Boeing? (Score:3, Informative)
Re:A Boeing? (Score:3, Funny)
Re:Then vs Than (Score:2)
Why not just overlook such a simple thing and judge the article by its subject matter? No need to go off on a rant and p
Re:Then vs Than (Score:2, Funny)
- A nitwit