Hi,
I'm working on a QT application that has several views / viewports onto a scene
(different perspectives and projections). One of these needs to have a
transparent QT HUD. I've spent a lot of time going over examples and archive
posts, and it seems to me that the most up-to-date way to do th
I've made a couple of attempts, and I feel like I am very close...
On one attempt, using a seperate osgViewer for each View, everything works but
the Database Pager isn't happy about having multiple viewers looking at the
same data. (osgEarth data under PagedLOD nodes flickers in and out of the
I'm using the new version, under Linux (although I will eventually be testing
under windows as well, I need it to work on Ubuntu first).
Like I mentioned, I am trying to extend the example to fit into the
QGraphicsView approach. Based on various guides and discussions that I have
found, this in
Hi,
I'm looking at extending the osgViewerQT example by being able to add items to
the QGraphicsView (by setting a QGraphicsScene and adding items to that). These
items include some QT controls that need to be embedded in the window with
transparency enabled.
When I simply create the scene and
Hi,
I'm currently walking down the oft-treaded path of creating Qt Widget overlays
on top of an OSG scene. I've been doing a lot of reading and searching over the
archives, and it seems like there are 2 main approaches. Eveything I can find
on the topic is 3-5 years old, however, so I was wonde
Hi,
Is there any way to specify the order that your event handlers / camera
manipulator will see events? In my application, under certain circumstances, I
want to override the default left-click behavior from camera manipulation to
something else.
Thank you!
Cheers,
Nick
--
R
Thanks Robert,
My branch is using 3.0.1, which did not have that functionality, but I was able
to piece it together from the implementation in 3.2.0, since it doesn't appear
that we are upgrading any time soon.
Thanks for the pointer-works just the way I hoped!
Thank you!
Cheers,
Nick
--
Is there a quick/convenient way to get the tex-coord for the vertex/drawable
returned from an intersection (i.e. the way that osgpick example returns the
vertex world and local coordinates)? Or do you have to compute the tex coord
yourself?
Cheers,
Nick
--
Read this topic onlin
I ended up solving 1 and 2 by using a single texture for the RTT to render the
whole window, then setting the drawable's tex coords such that each quad zoomed
in on the appropriate area. I haven't decided what to do about the resizing,
but it doesn't seem to be stretching the texture (which I wa
Hi,
I'm working on a zoom/magnifying lens feature, that I am implementing using
RTT. I'm setting up a slave camera to the main view and rendering to a texture,
which I place on a quad on the HUD. I've got a few questions:
1. Whats the best way to move the slave camera to magnify different areas
In case anyone comes across this, I ended up solving the problem by subclassing
the InteractiveImageHandler, QWidgetImage, and QGraphicsViewAdapter.
I had to get the InteractiveImageHandler to do the right coordinate projection
for the the clicks, and I had to get the QWidgetImage and QGraphicsV
I've managed to get the HUD to process events. The problem is that it is now
capturing all events, and I cannot use the camera manipulators underneath the
HUD.
It seems to be the case that the InteractiveWidgetHandler is always returning
true, because the widget is full-screen and thus always g
I found that I was using the wrong constructor for the Interactive Image
Handler, I think; It seems that the IIH expects the image to either be part of
the scene OR be a full screen HUD, not the sort of half-and-half that I have.
Using the other constructor for the IIH, one of my widgetImages ge
I have tried adding the InteractiveImageHandler to the view that I want to
handle them, but they still do not get any events.
Any other suggestions?
Cheers,
Nick
--
Read this topic online here:
http://forum.openscenegraph.org/viewtopic.php?p=59038#59038
I have a Composite Viewer setup with multiple views/cameras, and a 2D Ortho HUD
camera that I add to the root node of my scene to render HUD elements.
One of these elements is a textured quad with a QWidgetImage. I set up the
InteractiveImageHandler and gave it to the drawable as event and cull
Hi,
I tracked this down to a driver issue, I think;
I'm running on a laptop with Nvidia Optimus graphics using Ubuntu 12.04, and
support for the dual-gpu is a little wonky. I reconfigured my drivers, xorg
headers, and bumblebee optimus support, and got everything working happily.
Thank you! Mod
nsm4d wrote:
> Hi,
>
> I pulled the osgviewerQt example code into my project, and built and ran
> that. 3 out of 4 times, it runs fast (~60 fps), however, sometimes it runs
> slowly (~7 fps). So I don't think that the problem is my dataset or viewer
> setup. Attached are the stats when it runs
Hi,
I have a scene graph set up with a large dataset ( A Geocentric Earth from
osgEarth). In my composite viewer, I want to have multiple viewpoints (with
separate camera manipulators/projection matrices; ie. an Orthographic,
perspective, etc) each within a QT widget.
I got this set up the way
Hi,
I have a simple scene where I create a textured quad, and then add some simple
objects to it (spheres, lines, etc). Using the ubuntu distro OSG (3.0.0 I
believe), this was working fine. When I upgraded to 3.3.0, adding the new
objects now seems to overwrite the colors in the texture - for e
Hi,
When providing elevation data to osgdem. is it expected to be in HAE (height
above ellipoid), or MSL (mean sea level), or can it tell from the input? For
example, I am providing elevation data in the AIG/Arc/Info Binary Grid format
(.adf). The data represents the elevation of the area in m
20 matches
Mail list logo