Hi Peter,

I am intrigued. I plan to install OSM2World and follow it.

I work on the problematics from another side. I work on using real flights to 
capture data for 3D mapping and make it readily available for all mappers.

It is possible to publish 3D aerial images both into a Wikipedia article and a 
respective Wikipedia category. There's an OSM tag also for a category.

Here are the aerial images of Collège Madame de Staël in Geneva which I made 
with the DJI Phatom 3 this week:

https://fr.m.wikipedia.org/wiki/Coll%C3%A8ge_Madame_de_Sta%C3%ABl#

https://commons.m.wikimedia.org/wiki/User:Alexey_M./gallery

An advantage of Wikipedia is that it keeps the original images resolution, so 
building:levels are well visible. Google Maps automatically reduces resolution.

DJI Phantom 3 has got the range of 2.2 km, and the new Phantom 4 has got the 
range of 4.5 km (3 miles). But the range is also limited by the constant visual 
contact required by law. So actually it is about 1 km (and the maximum allowed 
altitude 150 meters in this case). Still it is enough to cover quite a large 
area.

Phantom 4 has got increased reliability due to doubled systems, but in any case 
it's recommended to perform hundreds of training flights before even 
considering a takeoff in a city.

DJI Phantom writes coordinates and altitude into JPGs Exif automatically, so an 
automation is possible.

Best regards
Oleksiy

Sent from my acer Liquid Z630On Mar 18, 2016 10:11 PM, Peter Barth wrote: > > 
Hi Wilson, > > Thank you for your interest in OSM and GSoC. > > The project 
idea you're talking about is actually meant in a > different way. What we'd 
like to achive is a way to generate > virtual camera flights. I.e. you use the 
OSM2World GUI to define > waypoints and direction vectors to look at, i.e. the 
definition > of the camera movement. The student's code would try to > 
interpolate those positions to generate a smooth trajectory in 3D > space and 
afterwards render the images along this > trajectory. Rendering might either 
happen directly with OSM2World > to output PNG images or via Povray. And in the 
end, ffmpeg/libav > will take these images and make a video out of the single 
images. > > Of course there's more to it: The trajectory might be inputed via > 
a GPX file or by defining your own format. Also the interpolation > of the 
camera movement might be as simple as a polygon or > something more 
sophisticated as bspline approximation or whatever > else. > > But in the end 
the details are up to the student. We have a rough > idea, we also have an idea 
how we would implement it. But we want > to hear and see your ideas as long as 
it matches our overall goal :) > > Hope that helps, > Peda > > 
_______________________________________________ > dev mailing list > 
[email protected] > https://lists.openstreetmap.org/listinfo/dev
_______________________________________________
dev mailing list
[email protected]
https://lists.openstreetmap.org/listinfo/dev

Reply via email to