hey Nitant,
afaik, this is not possible.
to be very honest, right now i cannot even think of any advantages in
your suggestion/idea.
if i understant you correctly, you're suggesting something comparable to
FBX or Alembic files
for deep pixel data. i don't think that you'd end up with a
substantially smaller file compared
to a frame sequence. in your example of a static object and a moving
camera, FBX and the likes
save a lot of space because e.g. an object is saved in it once plus the
camera animation. the
object is not saved per frame. now with deep pixel data, i highly doubt
that there's a lot of this
kind of space saving possible. so the only way to reduce size is
compression. to achieve a good
amount of compression it should be done across multiple (or even all)
frames. which means
every frame needs to be rendered first. so this would be a post-conversion.
even if this were a useful approach, post-processing a lot of rendered
frames into one
big file is probably not very reasonable.
next, re-rendering erroneous frames would be basically impossible or
result in the need to
re-build the all-frames-in-one-file from scratch.
what you're imagining is roughly comparable to a multilayer-exr or
stereo-exr. there is no
speed-up in using those compared to single layer/view images. for
example, to read layer 3,
the file reader needs to go through the file across layers 1 and 2 until
it can read the requested
layer 3 data. i admit, i don't know 100% how this works but apparently
there is a certain amount
of time wasted this way. that's why i'm a fan of using exr files per
pass instead of multilayer-exrs.
now imagine this issue being spread across a ton of frames instead of a
few layers and there's
probably a lot more file reading/seeking time wasted.
i'm not a programmer, so what i wrote might not be 100% correct. but my
'binary sense' somehow
tells me that i got the basics right ;)
so for what you'd like to achieve, i suggest going the FBX, OBJ sequence
or Alembic way if you need
to extract masks or similar tasks - in case you don't have those in the
rendered frames already anyway.
cheers,
Holger
Am 24.02.2013 10:22, schrieb Nitant Karnik:
Hey!
I was wondering if there was a way to write out one deep image file
per object for the entire length of a shot rather than a deep image
sequence... almost like a kind of DeepGeo?
Let's say the element you were rendering deep data for was a static
object like a building, light post, or a static cloud, couldn't you
pre-calculate all angles along the path of the camera for the duration
of the shot in one file? I would imagine it would almost be like in
comp terms, 'max merging' every deep frame together and rendering that
out?
I think ultimately having 1 file of deep data would be smaller and
less of a network hit then having 200 files of deep data with almost
50% (or more) of it being redundant information.
I'm sure someone's thought of this already... any luck implementing it?
Nitant
_______________________________________________
Nuke-users mailing list
[email protected], http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
_______________________________________________
Nuke-users mailing list
[email protected], http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users