Hi Matthias,  

You're a genius. Your method produces a spherical map of the whole scene. 

For color and brightness there are some hurdles to take though. Because the
illumination is evaluated instead of the color objects mapped with
illumination maps come out relatively bright while objects that recieve just
ambient light and that are not directly lit come out black. 
All this could be overcome by putting up light and textures in a scene
twice: once for ray traced images and once for the VR presentation.
Depending on the scene this may be a lot of work and result in two different
outcomes.

Thanks to the help on this list I now know what I did not know before: I
need a program that will stitch rows of prerendered images together and that
will turn them into something that a quicktime viewer understands. Probably
PTGui.

Thanks for your help,

                Frank Bueters


> I've simply created one big analtic sphere (2000units) and 
> there inside a small analtic sphere (1unit)
> 
> Then I've assigned the colorgrid material to the big sphere 
> and chrome material to the small sphere.
> 
> From chrome material I've deleted the
> specular shading
> 
> Then I've evaluated the illumination
> from the chrome sphere (see settings on the picture except 
> that I've used as size 1024x512)
> 
> I've converted the image with photoshop cs3 to jpg...
> 
> (the red sphere in the image was only a testsphere)
> 
> Matthias
> 
> (In short: no cam in scene, only illumination evaluated)

Reply via email to