Mathias Buhr wrote:
> Hi everyone,
> 
> I'm trying to use the stencil buffer to render only parts of the final
> image through other graph-attached cameras. The app works fine on ATI
> hardware but NVIDIA fails. OpenGL reports that the stencil buffer has
> only 0 bits (which is 8 bits on ATI, via
> glGetInterv(GL_STENCIL_BUFFER_BITS) in a DrawCallback). I have tested
> this on various linux boxes (Ubuntu 9.04, Fedora 11) with recent
> hardware and recent drivers.
> I'm pretty sure that a stencil buffer is available on this hardware
> because OSG seems to be able to utilize it (stereo-mode in
> osgUtil::SceneView uses stenciling and works fine). I'm pretty sure I've
> missed something. Is there anything special to do to get or enable a
> stencil buffer on Nvidia?

My OSG application uses a stencil buffer on Nvidia hardware, so it is
definitely possible. There should be a mechanism (depending on how you
create your context) to request a stencil-buffer-enabled context. I'm
using FLTK to create my OpenGL context, so my code's example might not
be useful to you.

It looks like osg/DisplaySettings has a method setMinimumNumStencilBits
that may be helpful to you. You could also look at the osgreflect demo.

-Eric
_______________________________________________
osg-users mailing list
[email protected]
http://lists.openscenegraph.org/listinfo.cgi/osg-users-openscenegraph.org

Reply via email to