See if RotoPDF works
http://greyangle.com/nuke/docs/geometry/RotoPDF.htm
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
No magic tricks just lots of different keys. ie soft key for hair, veil and
hard key for everything else. Throw some roto in places where it's easier than
messing with the keyer. Keylight works very well. Use a different de-spilling
setup and combine with the matte at the end.
Definitely try it out. Super fast and cool tools. Should be shipped with every
Nuke!
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
Preferences/Viewers/roto - Points
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
Unfortunately I dont have the files on me right now to create a setup for this
one. The theory is with a tracked camera you can 'project' a fix or new imagery
onto the original filmed plate, and it 'automatically' will apply the correct
perspective and movement changes.
Check out a tutorial
TheDagNode wrote:
But the question wtill remain, can you use the 2D track and calculate a
camera move based on that?
Yes, that's what the camera tracker does. It takes these 2d tracks to derive a
3d camera. In theory you can setup a full 3d camera track on user tracked 2d
tracks.
You can also write out a sequence of card randomly rotating on Y and use it as
a random input to the particle pipe of the particle emitter.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
It's a Z depth composite using a deep pipeline.
You might want to do this to visualise in 3d space a z composite or to
intergrate an element into a 'proper' deep composite.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk,
That's the correct behaviour of the Autocrop checkbox. Hover the mouse over it
to see it's usage.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
What does you ViewMetaData node display? If it's 'exr/compression = 2' then
it's Zip and '4' for Piz
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
Consider using SynthEyes ~ this took approx. 5-10mins.
Easytrack2.nk
Description: Binary data
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
Good call John.
Mantra render with 16bit 'C' 32bit extra planes = 2.5Mb
Mantra render with 16bit 'C' 16bit extra planes = 1Mb
Re-render in Nuke at 16bit = 760KB
In Mantra just set the Quantize of the Extra Image Planes to 16bit float from
32bit float.
It's just a workflow issue, not an hard rule. Otherwise we would be always
working with 6k+ raw-bayer images at all times wouldn't we ;)
You can always fix under/over-exposures, color-space convert, de-vignette,
de-noise, re-sample to 1/2k from 4/6k before keying etc if that's better for
the
works fine on os x.
launch nuke with --safe and test.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
Not sure if this works for your setup but for remote rendering on os X, I ssh
into the render box and launch nuke -x [script name]
Now the million $ question... Is there any way to successfully run Nuke
from the command terminal? I'm trying to initiate a render job from a
remote terminal,
Use Nuke's Scene node to set the display to 'off' or Atomkraft's
'AtomReadAsset' node that can load in the geo as points or bounding boxes for
viewport display and then render them fully at rendertime. Simply export the
instanced spheres as Alembic format and read them back in.
As the camera is jittering so much you may have to use 2d trackers or manual
tracking and use 'User Tracks' in the Camera Tracker, Tracking tab.
There also may not be enough parallax to do a 3d camera solve
___
Nuke-users mailing list
It might be easiest to simply set a grade for the shaded and sunlit frames and
then cross dissolve by eye or use the CurveTool to sample the actual pixel
values and then attach the values to a grade node.
___
Nuke-users mailing list
Try a Crop node after the transform
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
ReadGeo node/ uncheck 'read on each frame' / set keys with 'lock frame' and
you can edit the curve in the graph editor
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
that will work.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
http://support.thefoundry.co.uk/cgi-bin/mailman/listinfo/nuke-users
Yes - just put the image on a 'card' - paste the text below into the your node
graph to see an example.
set cut_paste_input [stack 0]
version 7.0 v6
CheckerBoard2 {
inputs 0
name CheckerBoard1
selected true
xpos -264
ypos -196
}
Card2 {
control_points {3 3 3 6
1 {-0.5 -0.5 0} 0
Like this?:
set cut_paste_input [stack 0]
version 7.0 v6
Camera2 {
inputs 0
translate {1.631346107 0.4529297352 2.102210045}
rotate {-9.659988623 37.81199213 0}
name Camera2
selected true
xpos -330
ypos 125
}
set N7340fb0 [stack 0]
Light2 {
inputs 0
translate {0.555943 0
You should be able to use a ViewMetaData node to see the colour space that the
dpx is scanned in with.
It might be Log but it could be something else.
___
Nuke-users mailing list
Nuke-users@support.thefoundry.co.uk, http://forums.thefoundry.co.uk/
24 matches
Mail list logo