Quake III Gets Real Time Ray-Tracing Treatment 116
Ozh writes "Did you ever wonder what you could do with a cluster of 20 AMD XP 1800s? Some German students and videogame fans did, and their answer has been what they call 'ray-tracing egoshooters', an entirely raytraced game engine which 'runs about 20 fps@36 GHz in 512x512 with 4xFSAA'. The first game to get this treatment is Quake 3 Arena : the screenshots look slightly better than the original 3D engine but the video (56 Mb, 3'19) is quite dramatic."
Does it make full use of GPUs? (Score:5, Interesting)
They do also mention that it can render entirely in software over the network at 20FPS - not bad considering that each fram portion of the data has to pass across presumable 2 machines before it is passed to the display!
1800Mhz x 20 = approx 36Ghz (Score:2)
Re:1800Mhz x 20 = approx 36Ghz (Score:2)
Well.. I'm not sure where the summary got it's specs of AMD XP 1800s, but they are not 1.8 GHz processors. They run at 1.533 GHz. That makes it 30.66 GHz (assuming the 20 computers). Looking at the site, there's absolutely no specs on the system aside from the 36 GHz number:
Re:1800Mhz x 20 = approx 36Ghz (Score:2, Informative)
> This is realtime speed for a virtual intel CPU with about 36 GHz (to be more precise: a cluster with 20 AMD XP1800 was used).
20xAMD compared to 1 intel?? weird comparison! (Score:2)
The only reason I can think of them consciously doing this is to up the speed numbers, since 30GHz is in our minds not as impressive as
Kinda cool (Score:5, Interesting)
They need to soften the shadows also. Either by using tricks or radiosity. Right now it looks kinda meh...
Interesting effort though.
Re:Kinda cool (Score:1)
Re:Kinda cool (Score:5, Informative)
Re:Kinda cool (Score:2)
Re:Kinda cool (Score:2)
Perhaps Google will have a Google free day, where for the whole day their 100K computers do Folding@HOME or something else other than google search.
Freecache link for the video (Score:3, Informative)
Re:Freecache link for the video (Score:4, Informative)
And besides, I'm _still_ getting over 100k/s (it was 115 when story first posted, now at 105) so we're collectively not doing the greatest job of slashdotting them anyways.
Re:Freecache link for the video (Score:2, Funny)
Nice links there (Score:2)
Screen Shots [uni-sb.de]
Downloads (video) [uni-sb.de]
raytracing downsides? (Score:4, Interesting)
So my question is, for those of us who don't know the first thing about 3D graphics, what are the pros and cons of a raytracing GPU, compared to the polygon pushers we currently know and love.
Re:raytracing downsides? (Score:5, Informative)
The real benefit is free occulsion culling, shadows, lighting, reflections, and essentially a physical simulation of how things actually work in life. There's been a few boards prototyped to do ray-tracing. Just google for Real-Time Raytracing. The paper behind it suggests that a hardware raytracer scales nearly linearly with the amount of tracer units behind it. These days its difficult to take a hardware prototype and beat the market standard with a wholly different paradigm, especially when the benchmark is OpenGL based. OpenGL only provides for 4 light sources, and little point. The prototype that exists is incredibly large and not well suited to current small desktop cases. But given the right set of talent, this is an interesting concept that could prove to take over poly pushers eventually.
Re:raytracing downsides? (Score:5, Informative)
I remember seeing an SGI demo of real-time raytracing. They used a kinda neat technique -- the application was a modelling app, and so they "faded" the new image in by rendering random pixels. It let them get away with far fewer FPS.
Re:raytracing downsides? (Score:5, Insightful)
A freind of mine wrote a real time raytracing engine as an assembly demo on an 80386 back in 1993 or so. Doing "real time raytracing" isn't that hard, it's doing it with complex objects, lots of light sources, and high resolution, that becomes a problem. IIRC, his was in 320x240, and was only rendering half the lines on the screen (effectively 160x240), and the scene was just moving through a dullish rocky martian landscape with a setting sun as the only lightsource.
My point being, it's not a great feat to do realtime raytracing, it's just a great feat to harness enough hardware power or come up with enough optimization tricks (without cheating and make it of lesser quality than a real raytrace) to do big nice-looking things with it.
Re:raytracing downsides? (Score:1)
Re:raytracing downsides? (Score:2)
Don't know if you're still reading this AC, but I'll go search around a bit if you bookmark this spot. The guy was a high school freind of mine, I doubt it was published in the demo scene, but I think he might still have a copy on a website of his.
Re:raytracing downsides? (Score:2)
Ok a couple minutes later here I did find a trace of it (pun intended). In '98 the guy made a crappy homepage, and one of the sections was on his "landscape raytracing". See web.archive.org url here (site is long dead): http://web.archive.org/web/20000420045611/www.lei d ich.com/raytrace.htm
The program itself isn't there on that site, but web.archive.org captured some of the resultant artwork. You'll notice he claims he generated them with a raytracer he wrote in assembler in less than 4000 bytes, and t
Re:raytracing downsides? (Score:5, Informative)
Umm.. no. In practice, no raytracer traces rays from the light source because very few of those rays would ever hit the camera. Instead, all raytracers do it backwards: backtrace the ray that would come from the top left pixel orientation towards the camera lens. When the ray hits an object (say, a wall), compute backtraces from that location. If you don't need realistic lightning, hitting a wall could always return preset amount of light (mixed with the object texture, of course) from that wall and no scattering of the ray.
The problem with full hardware raytracers is that the hardware should be able to hold whole scene or there'll be problems with some ray directions. GPU and the board on which it recides would limit the complexity of the scene, unlike with OpenGL which may render as complex scenes as the whole system can store (part of the scene can be streamed from the hard drive...)
I think the future will be a mix of both systems. Raytracer for curved, reflective surfaces. Multipass raster engine for everything else.
After looking through the video clip, it seems clear to me that the most important improvement in current games is better shadows. How many reflective surfaces there're in your environment? I'd say the glass is only one I'd miss reflections from and if that makes the difference between 2fps and 200fps, the lets forget the real reflections and use environment cube mapping instead.
Re:raytracing downsides? (Score:2)
Also, FWIW, there are more advanced (and expensive) forms of ray tracing that do involve tracing rays from light sources into the environment, to simula
Re:raytracing downsides? (Score:1)
Re:raytracing downsides? (Score:2)
Re:raytracing downsides? (Score:5, Informative)
Raytracing is generally more expensive than traditional polygon based graphics. You get more realistic curvature, far more realistic lighting, (including incidental light, diffuse light, etc), reflections, deflections / transparencies (such as those glass balls everyone loves), etc, etc, etc.
When Pixar goes rendering, Pixar raytraces. When Cameron goes rendering, Cameron raytraces.
The downside is that raytracing is a total resource hog. Essentially, for every pixel on the screen you trace the path of the light backwards, discovering every incidental surface and light source that might be effecting it along the way.
Polygon algorithims put stuff immediately to the screen, only going so far as to cull the faces that aren't visible to the camera. This is a lot more efficient for today's graphics, and will be far into the future.
And every time we get a step closer to using realtime raytracing, we get better polygon altorithims. First we had flat polygons, then we had colored vertexes, now we texture a character based upon averages of the normals of the surrounding vertexts, creating seamless skins. Originally we had no light, then a baked in faked lighting, now we have multiple light sources with multiple faked shadows on a baked environment. Glass and mirrors, once unheard of in a videogame, are now common. We even sample textures over a given area to try and get a more accurate per pixel representation.
So to answer your question, a raytracing GPU would have to be bloody powerful to do what you can do today with a polygon engine in realtime. Again, everyone thinks we'll get there someday, and there is no doubt in my mind that we will, but a realtime raytraced commercial game is such a distant possibility as to be a lifelong aspiration.
Re:raytracing downsides? (Score:3, Informative)
No, they don't. At least not when they can help it. Renderman didn't even have raytracing capabilities last time I looked, which admittedly was a while ago - for scenes that absolutely need it, "frankenrenders" using mental ray or BMRT were the order of the day.
You can do a hell of a lot with reflection mapping and custom shaders in a lot less time.
Re:raytracing downsides? (Score:1, Informative)
Re:raytracing downsides? (Score:1)
From what I understand of Renderman (haven't used it personally), I believe it breaks up your scene into polygons that are of subpixel size, rendering with that. A major attraction of it is its shader language, whose flexibility we are starting to see to some extent in hardware now.
Re:raytracing downsides? (Score:1, Informative)
Re:raytracing downsides? (Score:3, Informative)
> graphics, what are the pros and cons of a raytracing GPU, compared to the
> polygon pushers we currently know and love.
Raytracing requires a good deal more out of your hardware. They're running a
twenty-node cluster and only getting 20 fps, and I bet they're not even doing
some of the fancier tricks raytracing is capable of doing. So the downside is,
raytracing is slower, a *lot* slower on the same hardware.
The advantages
Re:raytracing downsides? (Score:2, Interesting)
Re:raytracing downsides? (Score:2, Interesting)
> 2^(20/1.5) times more powerful, which is to say 10321x more powerful than today
Yes, but meanwhile expectations (regarding resolution, object complexity, and so
on and so forth) also keep going up, as existing game technology develops. It
is not enough to have the processing power to raytrace something that would
look very impressive today.
20 years could be longer than is required, but it's very hard to tell ahead of
time how lon
10-15 fps (Score:1)
You were probably one of those masochistic bastards that installed Windows 95 on a 4MB 386SX, weren't you?
Re:10-15 fps (Score:2)
God that was awful. But after doing that, I made my father use the computer. He realized how slow it was, and two weeks later I had a spiffy 133 w/ 32mb ram.
Re:raytracing downsides? (Score:5, Informative)
1. Imagine a ray shooting out of your eye through each pixel on your screen. So ray shoots into the 3D world and it might hit an object. The purpose of the ray is to collect light information.
2. If it hits an object, it will bounce off at a certain angle (depending on the object). After it bounces around a couple times, it might eventually hit a light source or you might set a limit to how many times it can bounce. Each time it bounces off an object, it might lose some intensity depending on the surface of the object.
3. After all the bouncing, it collects light information (depending on what it hits, the surface, the lighting) and now that pixel now has more accurate light info for rendering.
What this allows is much more realistic mirrors, reflections, lightings, shadows, etc. but as you can imagine, bouncing off all those rays takes lots and lots of computation. Radiosity was mentioned, and that's basically shooting millions of rays FROM the light sources first (instead of from the users eye to each pixel). But again, lots of calculations.
If you're wondering how current games look so good without raytracing, it's due to lots of clever hacks and simulations for lighting/shading. Raytracing is kind of a brute-force/realistic method. Hope that helps someone...
Re:raytracing downsides? (Score:3, Informative)
Actually, that would be light tracing, which is rather different from radiosity. In radiosity, the scene is divided into small surfaces (patches) of approximately the same size, then the lighting distribution by diffuse interreflection is calculated. This does not have anything to do with ray tracing (ok, you can do it stoch
Ahhhh, yes. (Score:2)
For example:
Quake (1, 2 and 3)
Re:FSAA? (Score:2)
Probably because they're only rendering to 512x512.
Re:FSAA? (Score:5, Informative)
Re:FSAA? (Score:2, Troll)
They could be rendering it at nearly 1280x1024 at the same speed without AA... I'd take that over 512x512.
I've seen "games" that use realtime raytracing on ONE computer. If they need 20 computers to do the same thing, I have to wonder. Is their program just unoptimized/poorly coded? Or have they turned on many questionable quality gimmicks?
Re:FSAA? (Score:2)
Re:FSAA? (Score:2, Interesting)
A quick google search turned up http://www.realstorm.com/ which is a realtime raytracing engine. A very impressive one might I add. It is able to render at 512x512 in realtime on a single machine, and looks better than Q3RT to boot.
Re:FSAA? (Score:2)
Re:FSAA? (Score:2)
Re:FSAA? (Score:2)
Radiosity! (Score:2, Informative)
Re:Radiosity! (Score:5, Informative)
Radiosity would dramatically increase the computational complexity.
Polygonal rendering: O(N), where n=number of triangles
Ray tracing: O(log N), where n=number of objects (assuming a good bounding volume heirarchy)
Photon mapping: O(P log max(P, N)), where P=number of photons, which generally must be inserted into a kd-tree, and N=number of objects
Radiosity: O(N^2), where n=number of triangles
Ray tracing could conceivably make a game faster, if the scenes are complicated enough. Radiosity, on the other hand, is very very slow. Photon mapping [ucsd.edu] might be a better choice - it traces rays from the light source, and stores photons at the object intersection points, which are then used by the ray tracing step to approximate global illumination.
-jim
Re:Radiosity! (Score:2)
I was under the impression that ray tracing and radiosity weren't exclusive techniques, and using both produced excellent results.
ie ray tracing is excellent for reflections and refractions but tends to look too harsh when dealing with soft shadows and ambient light which is where radiosity works well.
Corrections welcome
Re:Radiosity! (Score:2)
Yes, radiosity and photon mapping are both usually used to calculate indirect lighting. Usually they're used in conjunction with ray tracing or polygonal rendering.
-jim
Re:Radiosity! (Score:2)
IMO, the best approach would be a hybrid of polygon+raytracing; use polygon rendering with shaders (and subdivision surfaces or some other form of arbitrary-precision surface tesselation) for most things, and then raytracing for reflections/refractions. There's a lot of things which pixel
You use radiosity... (Score:2)
Re:You use radiosity... (Score:2)
Where's the fun in that? I'd like to see dynamic global illumination (It might take awhile before computers are fast enough, and people figure out the right algorithms, though).
Photon mapping can be precomputed as well, and can simulate a number of effects that can't be simulated with radiosity (subsurface scattering, caustics off o
Re:Radiosity! (Score:1, Interesting)
Polygonal: O(max(S, N))
Ray tracing: O(S log N)
Photons: O(max(S, P) log max(P, N))
Radiosity: O(max(S, N^2))
Note that if N ~= S polygons keep linear behaviour while ray tracing becomes linearithmic.
Lan party upgrade (Score:4, Funny)
And I was just thinking about my next upgrade for HL2/Doom3.
Imagine a cluster... oh wait.. um, so is it running linux? and where is the source code?
Soon... (Score:4, Insightful)
Re:Soon... (Score:3, Informative)
20 computers running.
You have one computer.
log(20)/log(2) * 1.5 yrs = ~6.5 yrs
A bit longer, if you want a full 75 fps (or 60, if we're all using LCDs in 7 years).
Plus, technically Moore's law relates to transistor count, not processing power.
I'm interested in when we can do this [irtc.org] in a game in real-time. l(2hrs*3600secs/hr*60fps)/l(2) * 1.5yrs = 28 years before we see this in real-time (though that's using Pov-Ray, which could probably be sped up a lot if it's made into a game engine rather than a g
Re:Soon... (Score:2)
Re:Soon... (Score:2)
You have one computer.
log(20)/log(2) * 1.5 yrs = ~6.5 yrs
Yeah, but those 20 computers are only 1.8 GHz each. Desktops can currently go twice that fast, so take off 1.5 years or so from that total assuming that you go with the best commercially-available desktop technology.
Rob
Re:Soon... (Score:2)
Doom and Quake aren't bad, but they still leave a lot to be desired.
And don't get me started on how corny the sounds are...
Re:Soon... (Score:2)
[clip]
And don't get me started on how corny the sounds are...
Aureal tried commercializing "wave-tracing" at the consumer level via A3D -- actually tracing out the paths of sound waves.
Unfortunately, they got hijacked by the incumbent (Creative, who didn't have a lot of incentive to get involved in a research war) who introduced a small reverb model
Re:Soon... (Score:2)
I can't find it now, but I've seen at least one gradschool project on generating a new sound from an interaction, based on things like the interacting materials, the speed and location of the impact, etc etc. It was darn cool stuff.
-Zipwow
Re:Help? (Score:3, Informative)
And they did this for every pixel on each scene, 20 times per second.
It's a slow technique but it gives good results. They managed to do this fast by using hardware and 20 computers all running in parallel and transfering the results over the network in re
Re:Help? (Score:2)
Imagine a piece of paper. Shine a laser (or any light) at it. Assume the light and your eyes are a 90 degree rotation about the center of the point apart, and both looking down at the paper at an angle of 45 degrees. If you raytrace the rays from your eyes, they hit the paper, bou
Re:Help? (Score:2)
Most surfaces scatter the light, so you can bounce off in any direction chosen at random. You generally have to do quite a few bounces before you bounce into a light source or the sky.
Incidentally, perfectly parallel mirrors can be the worst case; imagine a beam of light at a small angle- how many times would it bounce between them before making it out the other end?
You are right though- generally a mixture of forwards and backwards techniques are used.
Re:Help? (Score:5, Informative)
Here is an example of a (not real time) raytraced image [irtc.org] (one that doesn't use radiosity -- just straight raytracing). In theory, given enough CPU power, they could pull this off.
Ray-tracing uses a method of 3d rendering that is currently beyond dedicated 3d hardware and must be done in software.
The main benefits of ray-tracing from a quality perspective are:
* True, accurate shadows from *everything* (most games, even stuff like Neverwinter Nights have hackish shadow engines that don't realistically display what lighting would look like in real life. These are calculated in real-time, not the precalculated shadows that you'll see in, say, Quake, where the light sources never move. You could throw a flickering lantern across a bar with bottles falling down and have all the bottles cast their own shadows.
* Advanced lighting. Currently, real-time 3d engines are very limited in the types of light they can produce -- generally, only spherical point sources of light.
* Refraction. You can have glass, ooze, or water truly refract light and distort images, not just use some sorta-lame effect to vaguely approximate it. Think of looking through a glass lens or a window in an old house.
* Volumetric fog (where you have "clumps" or "clouds" of fog, rather than just a global constant flag fog covering everything). Quake 3 had some rather (IMHO) impressive hacks to emulate volumetric fog -- ray tracing allows *true* volumetric fog -- people vanishing in swirling clouds of fog and mist and the like, not just a straight visibility dropoff.
* Reflections (there are a lot of hacks to approximate this off with existing 3d engines), but raytracers are *made* for this sort of thing.
* True curved and arbitrarily-shaped surfaces.
* Light projections (with shadowing and all that). They show a bit of this in the demo -- you could have, say, two people having a swordfight in a theater and the picture washing over them, or a scene in a church, with dusty light from the stained glass windows washing over the characters.
Basic ray-tracing does have some flaws. The shadows are sharp and hard -- sharper and harder than in real life. There are hacks to do soft shadows, but there isn't a particularly good an efficient way to pull them off.
It's hard to deal with things like laser beams or light beams coming out of a prism in ray-tracing. You need to do forward raytracing/photon mapping for this, which I suspect that they aren't doing.
Ray tracers tend to look a bit "eerie", for lack of a better word. They tend to leave shadowed areas very dark -- in real life, light will bounce around in corners and things a bit (even surfaces that don't look "reflective" to us will do so). So if I shine a flashlight, a raytracer will show a perfectly accurate cone of light (unlike existing 3d engines) that will spill properly over all surfaces. However, that cone of light will be a *cone* -- normally, when I shine a flashlight in a room, it lights up the entire side to some degree because of light bouncing off of objects.
There are some really nice things about ray tracers. They tend to parallelize really well, so you can theoretically put lots of computers together to do renders (as these folks did), or have lots of chips in parallel to theoretically make a custom piece of hardware.
Re:Help? (Score:5, Insightful)
That's what radiosity is for. Now a realtime radiosity package would be tres amazing.
Re:Help? (Score:1)
Re:Help? (Score:2, Informative)
Photon mapping basically is tracing light particles from the light source over multiple bounces (for indirect illumination). The resulting hit points and other light data are then stored in a structure and used later on for lighting.
Path tracing in a way is like normal ray tracing, but you spawn new rays even for all kinds of reflection (that is *lots* of rays, which is why it's damn slow) and calculate the resulting color by weighting and addin
Stop the Insanity (Score:5, Funny)
freecache (Score:2)
pulling down at about 200kb/s
Wow. (Score:1, Funny)
It seems my question is already answered.
Re:Wow. (Score:1, Informative)
What is featured in the article IS a beowulf cluster.
Looks worse to me (Score:3, Interesting)
Q3 isn't designed, let alone optimized, for raytracing, so that's not a major surprise, but I still expected an improvement, not a downgrade.
I think a custom demo is called for.
The tech sure is hella cool, though.
Re:Looks worse to me (Score:3, Interesting)
Re:Looks worse to me (Score:3, Insightful)
Re:Looks worse to me (Score:2)
Yes. The reason why is they turned off one of the texturing passes. Q3 uses an extra drawing pass with B&W textures to make 'shadows' on the walls. Why'd they do this? I imagine because of the lighting effects in the game. If a flash goes off, you don't want the shadowed area staying dark. Just don't draw that shadow texture, and you get nice brightly lit wall caused by a rocket. (I apologize if that doesn't make sense. For a cle
I have to say ... (Score:2)
I can't wait to see this technology in production.
Re:I have to say ... (Score:1)
ed2k link (Score:3, Informative)
Re: ed2k, try this (OT) (Score:2, Informative)
Submit a patch or a feature request [sourceforge.net]. You're capable enough from what I've read. If no one posts even a requests in such a forum, there's no way to reasonably expect the code to get changed.
Re: ed2k, try this (OT) (Score:2)
The site's still up? (Score:2)
Ray traced games on consoles (Score:2, Interesting)
Slow already? (Score:2, Informative)
http://mirror.openbarr.com/20040509_egoshooters_q
An interisting debate on Real Time Raytracing... (Score:2, Interesting)
alienware (Score:1)
Re:alienware (Score:1)
Missings some things (Score:4, Interesting)
I guess they limited the path of the ray they calculated so it bounced only two or three times off an object before they stopped calculating it. (If they stopped after one pass you wouldn't have seen those reflective glass balls like you did, which need multiple passes to look like they do).
I also miss colour bleeding on the surfaces. E.g. when you have - let's say - a white surface next to a red surface, some of the red will bleed on the white because light coming from the red surface will fall on the white surface and light it in a red hue. You would have seen this with a proper raytracing engine where the light bounces multiple times from an object and where the colour of the light is affected by the colour of the object.
I think those are the main reasons why the video doesn't look as realistic as i hoped for. (Then again how realistic is walking through a building where they have decorated the place with gruesome wallpaper taken from a horror movie and gigantic brains on mechanic spider legs walk around...
None of those things... (Score:2)
Because when a ray hits a surface, it only spawns a reflection ray and rays to any (point) light sources. Therefore shadows will have sharp edges, either a particular light at a particular point is visible or it is not.
Moreover, the contribution of non-light sources to a surfaces' illumination is not modeled (light from OTHER surfaces).
To accomplish this, once must create a large number of trial rays in a pattern around the reflection
So....what? (Score:1)
POVRAY (Score:2)
Raytracing101 (Score:1)
Raytracing overview. (a simple implementation)
Raytracing involves the casting of rays from the camera's eye, through a pixel in the screen. For each object in the scene (assuming no partitioning optimizations) we calculate whether the ray intersects the object. If it does, we determine the distance from the 'screen' to that object (consider it the Z depth) and save this for later. Once we have checked every
Re:GRADE:C minus - not very well traced. (Score:2)
Give them a break. As it is, it took them 20 cpus to make this work. Area lighting (you were calling it Diffuse lighting, area lighting is the cg term usually...) requires an array of point lights to make the soft shadows you want. Yep, they're more realistic, but they are an order of mangitude harder to compute
Re:GRADE:C minus - not very well traced. (Score:1)
first of all I called it a "diffuse light source", not "diffuse lighting" - a critical distinction between the source and the product. What's more "diffuse lighting" is a cg term too, albeit one you don't seem to know, and "area lighting" is subset of it. If you sel