Hey!

I was wondering if there was a way to write out one deep image file per
object for the entire length of a shot rather than a deep image sequence...
almost like a kind of DeepGeo?

Let's say the element you were rendering deep data for was a static object
like a building, light post, or a static cloud, couldn't you pre-calculate
all angles along the path of the camera for the duration of the shot in one
file?  I would imagine it would almost be like in comp terms, 'max merging'
every deep frame together and rendering that out?

I think ultimately having 1 file of deep data would be smaller and less of
a network hit then having 200 files of deep data with almost 50% (or more)
of it being redundant information.

I'm sure someone's thought of this already... any luck implementing it?

Nitant
_______________________________________________
Nuke-users mailing list
[email protected], http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users

Reply via email to