Hello,
I am using OSG for a robotics simulator [1] already a while. Now I want to
implement a "camera sensor", meaning the simulation of a physical camera
mounted on a robot.
I got it to work with Render to Texture. However I have the following problem:
The rendering of the normal graphics is detached from the physical simulation
speed. E.g. I can run the physics as fast as I can and render still only
25fps.
However, the camera sensor needs to be synchronized with the physical
simulation.
How can I render the scene explicitly into the texture indepently of the
viewer? At the moment the scene graph looks as follows:
Viewer with main cam
- Root
- HUD display (with another camera overlay to see the texture)
- Camera sensor (RRT)
- World ....
- World
- Shadow
- Scene
I have also problems with the shadow when the RRT camera is used. I read [2-3]
to use slave views, but did not success.
Any suggestions would be highly appreciated.
Cheers,
Georg
[1] http://robot.informatik.uni-leipzig.de/videos
[2] http://www.mail-archive.com/osg-
[email protected]/msg33512.html
[3] http://www.mail-archive.com/osg-
[email protected]/msg36229.html
--
---- Georg Martius, Tel: +49 177 6413311 -----
------- http://www.flexman.homeip.net ----------
_______________________________________________
osg-users mailing list
[email protected]
http://lists.openscenegraph.org/listinfo.cgi/osg-users-openscenegraph.org