ch.
> What's an example of what you are customizing between projects?
>
> On Mon, Apr 24, 2017 at 5:36 PM, Michael Garrett <michaeld...@gmail.com>
> wrote:
>
>> The general approach is to define NUKE_PATH for that project, for the
>> custom nuke environm
The general approach is to define NUKE_PATH for that project, for the
custom nuke environment (like the .nuke folder but on the network). Then
you have the custom init.py, menu.py and gizmos etc living there. For
example, the most basic way of doing it in .bash_profile on OSX:
export
I'll admit on my Mac laptop I do it the easy, not that versatile way and
just do this in my login .bash_profile, using 'export' for setting
environment variables and also setting up an alias to launch Nuke from a
shell. This is how it's described in the Nuke docs. If you want the
variables to
Yes it's more than a shame the 'e' parameter doesn't work. The other thing
that may help in this situation is to bake out Transform nodes for
individual tracks and string them together. I try to get out of the Tracker
node as soon as possible. For baked Transform nodes, the transform data
needs to
Oh yeah, it's a drag. From memory what I think I did was I set up the
procedural file path using the tokens in the export template. I had to
manually go into the comp and render any quicktimes for dailies, from the
rendered exr's, with my custom Write node which would have some TCL in the
File
Hey Gary,
Amira LogC is the same isn't it? It's possible Premiere is applying a LogC
to Rec709 LUT to roll the highlights off and get it looking presentable. I
would imagine from what you're saying that the footage looks blown out in
Nuke because it's not doing the additional s-curve in the above
nodes, to be honest.
>
> We'll be able to give a more definitive answer once we've got your script
> in through Support. I hope that is some help in the meantime.
>
> Regards,
> Lucy
>
> On Wed, Oct 26, 2016 at 4:28 PM, Michael Garrett <michaeld...@gmail.com>
> w
there is a vast performance difference in certain situations.
On 26 October 2016 at 11:28, Michael Garrett <michaeld...@gmail.com> wrote:
> Hi, just wondering if anyone has any suggestions as to why a BlinkScript
> may be working fine on a Quadro K1200 on a cheap Linux box, bu
Hi, just wondering if anyone has any suggestions as to why a BlinkScript
may be working fine on a Quadro K1200 on a cheap Linux box, but take hours
on a Quadro K600 and 600 (mixed environment) on Windows. Trying to remotely
troubleshoot this.
I'm getting a draw speed of about a second or two for
Hi, Nuke Studio question:
It seems there's no trim left/right keyboard shortcut for setting in and
out points, like say, in After Effects. Right now I'm using the razor
shortcut "c" then deleting the bit I want to trim. However this takes more
clicks and is a bit fiddly.
Thanks,
Michael
Ant's script, or something like it, really needs to be part of the UI. I've
seen the issue where Nuke comps are supplied as qt refs for the timeline
and they are 24fps. There is no way to get that to conform automatically on
import to a 23.98 timeline. If the comp is generated by Nuke Studio then
:32 AM, michaeld...@gmail.com (Michael Garrett) wrote:
>
>> Operationally Nuke wont concatenate any colour ops, only transforms.
>>>
>>
>> Er, are you sure about that? ;D
>>
>
> I guess it depends on the definition of "concatenate", but if you use it
&
display buffer with reduced colour depth ?!
>
>
> So Nuke might actually be rendering the Data to an 8bit buffer as part of
> the INPUT_PROCESS then applying the vLUT on top ?
>
>
>
>
> ------
> *From:* nuke-users-boun...@support.thefoundry.co.uk <
> nuke-u
Not sure if this is common knowledge, but I'm guessing that it is:
I was looking into a workflow where I set the shot CDL as an input process,
with the show lut as a viewer process on top of that.
Unfortunately, the input process and viewer process do not concatenate. I
end up with milky blacks
Note that I wrote the wrong path to the .bashrc for all users (Ubuntu
here). I meant to write /etc/bash.bashrc, not /etc/bash/bashrc
On 5 October 2016 at 18:40, Michael Garrett <michaeld...@gmail.com> wrote:
> In case others would like to know, I found the solution which was simply
e the full path or assign it to a variable
within the script. Pretty straightforward stuff...
Cheers,
Michael
On 30 September 2016 at 10:41, Michael Garrett <michaeld...@gmail.com>
wrote:
> Hi Chris,
>
> Thanks, I've checked the first link before (I'm on Linux). I already tried
> a
might be worth a read:
>
> http://stackoverflow.com/questions/1615877/why-aliases-in-a-
> non-interactive-bash-shell-do-not-work
>
> http://superuser.com/questions/386345/what-is-the-correct-
> way-to-alias-applications-in-os-x-through-bash
>
> bests
> chris
>
>
>
Hi Charlie,
Thanks, you reminded me I did try sourcing, of course that was the
"include" equivalent I was thinking of. It didn't work for this
scenario, but I'll take another look.
Michael
On Friday, September 30, 2016, Carlos Trijueque
wrote:
> I can't test it right
Does anyone have a solution they can share about how to get an alias
working in a bash script? In unix nomenclature, I mean an alias to a binary
(such as nuke, ffmpeg, whatever) that will execute in a non-interactive
non-login shell context.
My unix skills are not the greatest, and googling has
I just saw your second reply. Duly noted...
On 26 September 2016 at 16:43, Michael Garrett <michaeld...@gmail.com>
wrote:
> Deke,
>
> Thanks for letting me know. Although I'm all for OCIO, I used a
> Vectorfield because I wanted to change the input colour space to log
>
gt;> Nuke's new flipbook is based on the timeline/hiero viewer. It is 100%
>> OCIO based. So it does not recognize any Nuke IP or viewer processes which
>> are not in OCIO.
>>
>> On Thu, Sep 22, 2016 at 1:59 PM, Michael Garrett <michaeld...@gmail.com>
>> wr
OK, thanks for confirming you have this issue too.
On 25 September 2016 at 07:42, Igor Majdandzic <subscripti...@badgerfx.com>
wrote:
> Yeah, Ive seen this with provided LUTs... haven't found a fix though.
>
> Am 22.09.2016 um 22:59 schrieb Michael Garrett:
>
> I just regis
I just registered a viewerProcess that is just a Vectorfield exported as a
gizmo, but I've found it does not get picked up when I try to use Nuke's
native flipbook. This is in 10.0v1, I will upgrade shortly to 10.0v4.
I am able to select the LUT from the dropdown in the flipbook options, but
when
segments. Thanks for bringing that up!
On 9 August 2016 at 20:58, Marten Blumen <mar...@gmail.com> wrote:
> Smart Vector has always worked but only used on < 250 frame IIRC.
>
> On 9 August 2016 at 06:44, Michael Garrett <michaeld...@gmail.com> wrote:
>
>> I'm getting
I'm getting an error when using VectorDistort with pre-rendered smart
vectors, stating "reference frame is outside the valid frame range", when
it isn't. It seems pretty arbitrary on what frame it decides to error out
on. Has anyone else seen this and know of a fix?
This is using Nuke 10.0v3
Fantastic! Can this be used as a general voxel render for anything you
would put into PositionToPoints? For a while I've been interested in a way
of being able to render a ptex type of texture atlas in Nuke, where it has
a corresponding world position location.
On 5 August 2016 at 05:48, Mike
; effects)
>
> *Howard **Jones*
> Visual Effects Supervisor
> m: 07973 265624 | e: how...@axis-vfx.com | w: www.axis-vfx.com
>
> On 12 Jul 2016, at 16:23, Michael Garrett <michaeld...@gmail.com> wrote:
>
> For me, SmartVector has been a game changer. The updated colo
For me, SmartVector has been a game changer. The updated colour management
is also welcome.
On 12 July 2016 at 10:22, Rich Bobo wrote:
> Hi all,
>
> Just wondering what is the general opinion on upgrading to Nuke 10? Since
> it’s been out for a while now, are there any major
Sorry for the OT, but has anyone else had problems with actually buying RV?
It seems maybe the issue is something up with tickets in their support
system not getting re-routed to the individual's email. I have a ticket
with them that has been open for months from when I had a trial, and now I
want
.
Thanks,
Michael
On 15 June 2016 at 22:16, Michael Garrett <michaeld...@gmail.com> wrote:
> Thanks Peter, I just tried this out and it seemed to work. I was having a
> bit of trouble adding the Avid video levels back in though. For some reason
> adding a Gade node and changing the bl
e the mov64 vs mov32 issue that popped up
>> before with certain codecs - have you tried switching the decoder setting?
>>
>> If I recall, mov64 is the default but still has some lingering broken
>> bits (prob related to ffmpeg, who knows...)
>>
>> On Wed, Jun 8,
*Gade = Grade! (apologies for the typo)
On 15 June 2016 at 22:16, Michael Garrett <michaeld...@gmail.com> wrote:
> Thanks Peter, I just tried this out and it seemed to work. I was having a
> bit of trouble adding the Avid video levels back in though. For some reason
> add
hat has no issues
> with the flicker, well as far as I could tell.
>
> regards
> -adam
>
>
> On 09/06/2016, at 12:32 PM, Michael Garrett <michaeld...@gmail.com> wrote:
>
> So based on your experience, I could just batch process with the Nuke
> python module on the command line
the node graph
> and transcode only one movie at a time.
>
> regards
> -adam
>
>
> On 08/06/2016, at 8:26 AM, Michael Garrett <michaeld...@gmail.com> wrote:
>
> I've actually managed to get rid of the flicker by using the exact drop
> frame rate of 23.97602
<info.mail@mac.com> wrote:
> i answer to your question..
>
> Yes I bug has been reported regarding this.
>
> if you need to work in nuke / convert footage last version that did not
> have this issue was 9.0v1
>
> regards
> -adam
>
>
> On 08/06/2
Has anyone else experienced flickering Avid quicktimes in Nuke? I've had
this issue in Nuke 9 and 10 on Mac and Linux. The only solution seems to be
to transcode to frames beforehand in After Effects. Actually, a linux-based
method would be handy to know.
Thanks,
Michael
gravity. Then play with the knobs until you are happy.
>
> Or do you have any specific question?
>
> Am 26.05.2016 um 22:37 schrieb Michael Garrett:
>
> I have not as yet used Nuke particles in production, but now seems to
> finally be the time. I have about ten shots that need falling/fl
I have not as yet used Nuke particles in production, but now seems to
finally be the time. I have about ten shots that need falling/floating grey
ash in a room, with a bit of turbulence. Somewhat like grey snow. The look
is established already with an element on a card which is some dust blowing
Has anyone come up with a nice tool for this yet? I notice that the matrix
in a 2d node is not concatenating when stacking 2d transforms. So I would,
roughly speaking, do something like link the 2d node's matrix to an Axis
node, then do the concatenation with multiple Axis nodes and plug the
I might be repurposing a render node with Nuke Assist, also floating
license RLM server, so I hope it's not a massive pain!
On 26 April 2016 at 21:00, Frank Rueter|OHUfx wrote:
> Hi all,
>
> I am seeing massive license troubles when using NukeAssist in such a way
> that the
Still pretty new to NS here. I have an updated aaf coming from Avid
(because when is the edit ever locked?) that will be conformed in NS and I
want to take my existing comps across from the previous conform where
applicable.
I gather that the workflow is to put the rendered comps on a new track
I forgot to add that in this scenario I would do a deep recolour with the
UV remapped image onto the deep opacity it was extracted from, and then I
get the transparency/AA back.
On 17 February 2016 at 19:20, Michael Garrett <michaeld...@gmail.com> wrote:
> Thanks for bringing up dee
Gary,
Try this old gizmo of mine, it may not be functional any more (aka. needs
updating...) so email me if you can't get it working:
http://www.nukepedia.com/gizmos/other/envrelight
It uses the camera, position and normal data like the Relight node, with
Latlong or Mirrorball image inputs.
I see this too and had not noticed it until that long ago. I assumed it was
because I use Localise, but maybe not based on what you're saying.
On 12 January 2016 at 07:13, Simon Björk wrote:
> I've been seeing this terminal message when working lately. It happens in
>
Michael,
>> http://malditochroma.com/about-2 might do 3D tracking as well.
>>
>>
>>
>> Ron Ganbar
>> email: ron...@gmail.com
>> tel: +44 (0)7968 007 309 [UK]
>> +972 (0)54 255 9765 [Israel]
>> url: http://ronganbar.wordpress.c
gt;>
>> nvidia has cuda downloads for ubuntu 14.04 and 15.04 on their developer
>> page. If I remember right, Mint 17 is based on ubuntu 14.04.
>>
>> https://developer.nvidia.com/cuda-downloads
>>
>> On Thu, Jan 7, 2016 at 7:51 PM, Michael Garrett <michaeld...
u just
> run a simple installation command. The set up an installation instructions
> are all very clear and concise.
>
> From here nuke sees the GPU and runs flawlessly. They have a specific
> .deb package for Ubuntu / mint users...
>
>
>
> Check it out. :)
>
&g
In the spirit of the canvassing for opinions on roto outsourcing, I'm
looking to know more about 3d tracking outsourcing options.
Not sure if I can get around the NDA issue on the current project, but
worth inquiring.
MPC Bangalore doesn't count ;)
Cheers,
Michael
Hey all,
I just got a new Dell workstation preinstalled with Ubuntu and am running
Nuke 9.0v8. I'm finding the UI is not very responsive and the GPU is not
recognised by Nuke in the prefs - I'm assuming the UI drag is connected to
the GPU. The GPU is a Quadro K1200 and I am running the latest
he Nvidia-drivers pre-installed? Or does it
> come with nouveau?
>
> If the latter, install the drivers ;)
>
> Cheers,
> Elias Ericsson
> Den 7 jan 2016 8:16 em skrev "Michael Garrett" <michaeld...@gmail.com>:
>
>> Hey all,
>>
>> I just got a n
I can third that book!
On 14 December 2015 at 17:32, Randy Little wrote:
> Not whoops CONFIRMATION. :-D Always nice to know someone else agrees
> with something.
>
> Randy S. Little
> http://www.rslittle.com/
> http://www.imdb.com/name/nm2325729/
>
>
>
> On Mon, Dec
That worked...thanks!
On 1 November 2015 at 16:56, Frank Rueter|OHUfx <fr...@ohufx.com> wrote:
> I haven't tried but you should be able to do this on any resolved 3D point
> (green points).
>
>
> On 1/11/15 2:50 am, Michael Garrett wrote:
>
> Thank you for pointing o
.com>
> fr...@ohufx.com> wrote:
>
> Are we talking about the same thing?
> When I camera track something here and solve the tracks, I can right click
> on any solved 3D point and create objects on it:
>
>
>
>
> On 31/10/15 7:04 am, Michael Garrett wrote:
>
I've been using the CameraTracker more these days and I noticed that when I
feed in 2D user tracks, after the track Nuke projects them into 3D space
which is potentially useful for generating an Axis.
In the user track tab, the xyz position is available like in the older
ProjectionSolver node,
It comes from the culture of colourists.
On 17 July 2015 at 17:28, Simon Blackledge
simon.blackle...@spacedigital.co.uk wrote:
What's wrong with Resolve?
Sent from my iPhone
On 17 Jul 2015, at 21:40, Michael Garrett michaeld...@gmail.com wrote:
There is definitely latitude, which
.
Gary Jaeger // Core Studio
249 Princeton Avenue
Half Moon Bay, CA 94019
650.728.7957 (direct)
650.728.7060 (main)
http://corestudio.com
On Jul 16, 2015, at 8:31 PM, Michael Garrett michaeld...@gmail.com
wrote:
Gary,
The basics of the ACES route through Resolve are:
- Set up the project
to use ACES or YRGB? or ACES log? Argh!
Gary Jaeger // Core Studio
249 Princeton Avenue
Half Moon Bay, CA 94019
650.728.7957 (direct)
650.728.7060 (main)
http://corestudio.com
On Jul 13, 2015, at 6:33 PM, Michael Garrett michaeld...@gmail.com
wrote:
needing to disable the ODT
Motion estimation on hard cuts can be a nice part of the recipe. This can
create some good tearing morphy transition effects.
There was that Kanye West video from a few years ago that used it really
effectively, also a Chairlift one.
Is there some way of getting a rendered movie output of the
Any comment on using ACEScg in Nuke, and why there is no Rec709 rrt? From
reading the ACEScg white paper, it does seem there is just a matrix that
can be used. Also I notice OCIO ColorSpace has ACES to rrt Rec709.
On 6 July 2015 at 13:55, Deke Kincaid d...@thefoundry.co.uk wrote:
Just a note to
.
On Mon, Jul 6, 2015 at 11:08 AM, Michael Garrett michaeld...@gmail.com
wrote:
Any comment on using ACEScg in Nuke, and why there is no Rec709 rrt? From
reading the ACEScg white paper, it does seem there is just a matrix that
can be used. Also I notice OCIO ColorSpace has ACES to rrt Rec709
Not sure if this is exactly what you're intending, but since we got flat
exr renders at the same time we would copy the bbox across to deep. I also
generated a z crop (not just x+y). All of this definitely helped.
On 3 June 2015 at 10:27, Thorsten Kaufmann
thorsten.kaufm...@mackevision.com
Word!
On 28 May 2015 at 00:54, Ben Pierre benpie...@gmail.com wrote:
I like it.
-- -
pixel:muncher
On May 27, 2015, at 11:27 PM, Michael Garrett michaeld...@gmail.com
wrote:
Sorry for the late reply. Did you find a solution? I misunderstood what
you were describing. The expression
of the sphere. And if you use multisamples with a
moving camera, then all samples are just crazy.
Do you have a exemple of that deep expression?
'Thanks
Sent from my iPad
On Apr 23, 2015, at 8:44 PM, Michael Garrett michaeld...@gmail.com
wrote:
Ooh I didn't know about those nodes...thanks
Launch it from a shell, and feed it multiple clips. Agreed, we need these
two things. Then intuitive keyboard shortcuts to either toggle/cycle
through multiple input clips or do a wipe (between two clips).
The syntax is important too:
nukeflip -c path/to/framesA.%04d.exr path/to/framesB.%04d.exr
DeepToPoints post Scanline.
There are hidden DeepDeOverlap, DeepOmit Deep Clip Z nodes available
after updating that may try to do something you're after.
On 22 April 2015 at 13:29, Michael Garrett michaeld...@gmail.com wrote:
I don't have Nuke open, but is there not a drop zero opacity
I don't have Nuke open, but is there not a drop zero opacity samples
checkbox on the ScanlineRender deep tab?
If not, the workaround that has worked for me is to run an expression with
DeepExpression that pushes the location of any deep samples with zero alpha
behind the camera frustum, then use
Us the IBK? It's the same kind of concept (screen subtraction, whatever).
On 3 March 2015 at 17:54, John Mangia j...@johnmangia.com wrote:
Key it in log maybe or grade it before the key? Not sure how much
*difference* it would make though...
On Tue, Mar 3, 2015 at 5:39 PM, Randy Little
Hi Jason,
From what I understand of what you're saying, I've done this exact thing in
the past using the ProjectionSolver. That's now been replaced by the
updated CameraTracker, although I haven't checked the workflow. Of course
you can still access the ProjectionSolver if need be.
It sounds
Have a look at this:
http://www.nukepedia.com/gizmos/3d/alembicreconstructor
On 10 November 2014 20:16, Ari Rubenstein a...@curvstudios.com wrote:
Is the sun itself animated or just the camera ?
Sent from my iPhone
On Nov 10, 2014, at 6:48 PM, John Coldrick john.coldr...@gmail.com
wrote:
Hey Gary,
Yes it definitely helps. You need to use pgBokeh to get a nice deep
defocus. As far as motion blur, the ScanlineRender node outputs deep motion
vectors but so far there is no deep VectorBlur node. Ideally these things
would need to be calculated at the same time to get around the old
/14 07:49, Michael Garrett wrote:
I'm using SphericalTransform in Nuke 8.0v5 on Windows 7 and it's extremely
slow - much slower than I've experienced on other platforms with prior
versions of Nuke. Is this a bug with the Windows build, or is it a larger
issue with Nuke 8.x?
Thanks,
Michael
back.
-Nathan
*From:* Michael Garrett michaeld...@gmail.com
*Sent:* Monday, October 13, 2014 9:41 AM
*To:* Nuke user discussion nuke-users@support.thefoundry.co.uk
*Subject:* Re: [Nuke-users] SphericalTransform slowness
In this case I needed the rotation knobs in SphericalTransform
I had a chance to check this on a Mac, and it's practically interactive for
a 4k latlong on an iMac in Nuke 7 or 8. This is compared to a recent dual
processor Xeon Windows workstation where it is pretty much unusable.
On 14 October 2014 12:07, Michael Garrett michaeld...@gmail.com wrote
of SphericalTransform here.
http://www.nukepedia.com/blink/transform/environmenttransform
Might help.
HP
On Sun, Oct 12, 2014 at 11:49 AM, Michael Garrett michaeld...@gmail.com
wrote:
I'm using SphericalTransform in Nuke 8.0v5 on Windows 7 and it's
extremely slow - much slower than I've experienced
We had this exact problem as described and it affected us quite a lot,
hosing our scripts. I figured it was the existing long channel names issue,
and due to different exr versions in the deep vs regular readers.
I think the fix was to open the script in a text editor and delete the
extra
.
Cheers,
r.
Rangi Sutton
VFX Supervisor
Cutting Edge
On 9 September 2014 23:32, Michael Garrett michaeld...@gmail.com
wrote:
We had this exact problem as described and it affected us quite a lot,
hosing our scripts. I figured it was the existing long channel names issue,
and due
,
HP
On Mon, Aug 25, 2014 at 3:11 PM, Michael Garrett michaeld...@gmail.com
wrote:
In fact I took maths for the angular map transform (latlong to angular, I
think) from an old Debevec Siggraph paper or something, or maybe the book
he wroteso maybe that was yours too! Since I know you
Are you on Windows?
On 27 August 2014 11:36, John Mangia j...@johnmangia.com wrote:
This has been an issue for years and I'm surprised it hasn't been
addressed. Nuke constantly sucks up all available ram, caching
anything that was viewed and not releasing it from memory. All memory
2014 12:03, John Mangia j...@johnmangia.com wrote:
Nope, Linux for years and at different facilities.
On Wed, Aug 27, 2014 at 12:00 PM, Michael Garrett michaeld...@gmail.com
wrote:
Are you on Windows?
On 27 August 2014 11:36, John Mangia j...@johnmangia.com wrote:
This has been
I may be wrong here, but I've actually seen something like this bug with
some operations that require heavy filtering, I seem to remember it getting
introduced with Nuke 8. I submitted a bug report to The Foundry with speed
comparisons with Nuke 7.
I first noticed it with STMap then
, Frank Rueter|OHUfx fr...@ohufx.com wrote:
damn, that leaves me dead in the water.
oh well, I guess I will just go to bed and be grateful for any frame I see
in the morning
On 27/08/14 01:31, Michael Garrett wrote:
I may be wrong here, but I've actually seen something like this bug with
some
Wow, ok
On 26 August 2014 17:12, michael vorberg pingkin...@googlemail.com wrote:
http://variety.com/2014/biz/news/google-buys-visual-effects-firm-zync-1201290967/
google bought Zync
2014-08-18 11:52 GMT+02:00 Julik Tarkhanov ju...@hecticelectric.nl:
It was very promising and it is
are needed?
HP
On Sun, Aug 24, 2014 at 7:07 PM, Michael Garrett michaeld...@gmail.com
wrote:
I put it together before I knew about that 0.5 pixel offset trick, I
would say it needs updating but then Haarm-Pieter Duiker's blink
EnvironmentBlur node looks like the real deal.
http
Definitely possible. I've seen a couple of UV map Blink things already but
haven't actually done anything myself yet. For example:
http://nickdeboar.wordpress.com/2014/01/02/nuke-8-blink-script/
I have the UV map math for some of the transforms and it could be applied
to Blink for sure. Have a
it some more thought tonight.
cheers,
Patrick
On 24.08.2014, at 15:51, Michael Garrett michaeld...@gmail.com wrote:
Definitely possible. I've seen a couple of UV map Blink things already but
haven't actually done anything myself yet. For example:
http://nickdeboar.wordpress.com/2014/01/02
Can you recommend a particular photogrammetry-based tool that creates dense
point clouds? Obviously there is Nuke's internal toolset which has provided
great results on occasion and I haven't fully exploited it yet.
On 30 June 2014 04:42, David Yu dave...@gmail.com wrote:
Sense and all the
I think I had not paid attention to it before because I thought it was one
of those old win-only programs. Great that it's on all platforms.
On 30 June 2014 16:40, theodor groeneboom plateryt...@gmail.com wrote:
Runs on win,osx linux too.
On 30 June 2014 21:35, Michael Garrett michaeld
I would set an expression based on anything above a very large number going
to zero. As long as that large number doesn't encompass any actual valid
scene values. Then you can run your MaxLuma curve tool.
depth.Z100?0:depth.Z
On 23 June 2014 12:24, HSK charhar...@gmail.com wrote:
Hi
use *DeepRecolor* to achieve
this... ?
Colin
On 18 June 2014 05:33, Michael Garrett michaeld...@gmail.com wrote:
Sounds like you're sorted, but I found this note on the basic method I
used. So you could use a flat bezier mask as you say, that is promoted to
deep then used for the deep
If all the samples are at the same depth in both A and B, easy, the
output samples are a simple mix between each sample of A and B. This is
the case for, say, keymixing in a DeepColorCorrect (where only the
existing sample values will be changed)
I managed to get a deep volumetric keymix
Yes that would be really useful, as long as we are still limited to working
in a single multichannel stream for deep images in Nuke.
On 17 June 2014 01:54, Frank Rueter|OHUfx fr...@ohufx.com wrote:
Hi all,
since we still don't have a DeepCopy node in the default Nuke dist, does
anybody
the soft part of the mask in flat image space).
Hopefully we are good to go now.
On 18/06/14 3:25 am, Michael Garrett wrote:
If all the samples are at the same depth in both A and B, easy, the
output samples are a simple mix between each sample of A and B. This is
the case for, say, keymixing
I think Resolve (and Resolve Lite) v10 uses ACES .2, but it has an LMT to
compensate for the differences with v9 (that used ACES 0.1) so that you can
match the look of v9.
On 6 March 2014 18:16, Frank Rueter fr...@beingfrank.info wrote:
haha, when i check Resolve Lite 10, it seems to be using
Very cool, and the Mantra to Nuke tool looks interesting too.
On 29 January 2014 10:19, Neil Rögnvaldr Scholes n...@uvfilms.co.ukwrote:
This looks awesome. Downloading now.
I checked out your website Theodore and the Mantra tools also look really
cool.
Neil Rögnvaldr Scholes
Is it feasible to use a lens impulse response then run through a
Convolve? Like LensKernelFFT on Nukepedia. It seems that it is able to
capture a complex characteristic like that.
It's a bit more awkward than shooting grids and using STMap per-channel
though.
On 21 January 2014 06:16, Julik
Really organic looking!
On 10 January 2014 02:47, Frank Rueter fr...@beingfrank.info wrote:
nice one.
I really like how you are utilising the OSCs for working on the image!!
On 10/01/14 20:37, Mike Wong | ax.gmail wrote:
hi guys,
we have just released a new little plug-in, hope
+1
On 1 January 2014 15:30, Howard Jones mrhowardjo...@yahoo.com wrote:
I'd be up for the option. The new colour ui does nothing for me.
Howard
On 1 Jan 2014, at 11:44, Diogo Girondi diogogiro...@gmail.com wrote:
Hi,
First of all, Happy New Year! I wish all an incredible 2014.
Just wanted to say, I see having a DeepVectorBlur as a potentially great
way of minimising deep image file size since there would be far less deep
samples stored. I'm finding that even for fully opaque non-volumetric
surfaces that I know will never intersect with anything, there is sample
bloat
This is actually something you can do with Vray in Nuke, at least with a
recorded demo I saw online. It would be great if we could just render it
through the ScanlineRender node though.
On 11 October 2013 19:23, Jonathan Egstad jegs...@earthlink.net wrote:
Whoops, I mean Darren… see, it has
Normally I would say you can use the world_matrix knob of the TransformGeo
to get the concatenated position (using cels 3, 7 and 11) but it depends on
whether the geometry itself already has a baked in offset or not - meaning
is it reading in at the origin or not.
If it's not animated and you
1 - 100 of 196 matches
Mail list logo