Hi Christian, On 26 November 2014 at 14:34, Christian Buchner <[email protected] > wrote:
> > The way this works is that the nVidia driver (OpenGL or DirectX) is > looking for this export symbol in the binary that it's linked against. > Whenever it is found, the driver prefers the nVidia graphics card over > Intel. > > I do not think that the binary file in question (i.e. the OSG library, or > the program using OSG) is able to control at run time whether or not this > export symbol will be visible to the driver. > If it's our application/library providing the global that the driver queries for, rather than us getting a setting a global that the driver provides then it simplifies things significantly for us - we just have to create the global variable via an extern and appropriate export symbol and the driver will do the rest. This will also mean it's safe to implement on non NVidia systems and should be possible to place control for it in osg::DisplaySettings. Robert.
_______________________________________________ osg-submissions mailing list [email protected] http://lists.openscenegraph.org/listinfo.cgi/osg-submissions-openscenegraph.org
