Hey! I was wondering if there was a way to write out one deep image file per object for the entire length of a shot rather than a deep image sequence... almost like a kind of DeepGeo?
Let's say the element you were rendering deep data for was a static object like a building, light post, or a static cloud, couldn't you pre-calculate all angles along the path of the camera for the duration of the shot in one file? I would imagine it would almost be like in comp terms, 'max merging' every deep frame together and rendering that out? I think ultimately having 1 file of deep data would be smaller and less of a network hit then having 200 files of deep data with almost 50% (or more) of it being redundant information. I'm sure someone's thought of this already... any luck implementing it? Nitant
_______________________________________________ Nuke-users mailing list [email protected], http://forums.thefoundry.co.uk/ http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
