Thanks!
I use it J
From: softimage-boun...@listproc.autodesk.com
[mailto:softimage-boun...@listproc.autodesk.com] On Behalf Of Petr Zloty
Sent: Tuesday, March 12, 2013 4:15 PM
To: softimage@listproc.autodesk.com
Subject: Re: Copy UV to cloned copies
Hi, quick way:
## Python script
Hi,
I´ve two machines, one xp64, one win7. both run xsi2012sap and both have
identical
mi-raysatsi2012_3_9_1_44 [portnumber]/tcp entries in their ../etc/services file.
Both machines check out a GUI/network license correctly, e.g. will start xsi
locally just fine
and properly access a shared
Update:
The crash is reproducable when trying to use Unified Sampling
for both cases, with and without the setenv.bat entry
set MI_FORCE_OLD_BSP=1
Thanks for having me waste a day´s work on this, Autodesk.
Cheers,
tim
P.S: And a honest Thank you to Stephen Blair for at least getting me
That video is quite impressive, however when I work, I'd need rather
immediate response to my changes to shader, texture, whatever. I want to
look the model in different angle in realtime. That is what Maya offers,
and what Max offers too. At this moment HQV is a bit choppy, I have to
flip the
well not only for game artist... after all viewport is our main window into
application and work and 100% of time you are in viewport. that should say
enough. if window is all foggy.. then no matter what great stuff you have
behind it it is of smaller use when you can;t properly use it :) a bit
Hi Leonard,
Thanks a lot for putting this tool together, I had a look at your 7
tutorial and it does look like a very handy tool.
In terms of how you distribute it, is something you need to figure out but
would definitely recommend this purchase to my company.
Even though there are quite a few
Just a side comment, my colleague next to me is working with max. He
just start every scene with a basic light setup, shadows on, AO on,
soft lighting, etc, and has a good approximation what's going on.
Unfortunately I can't do that. Hopefully, the next release will have
improvement, and the next
But, the HQV can support any renderer not just MRay, provided there are
shaders available. I don't see how they could reasonably be expected to
have out of the box support for 3rd party shaders that don't come bundled
with the package...
On Wed, Mar 13, 2013 at 12:36 PM, Mirko Jankovic
Hello friends,
I have very small doubt about how to use multiple Filter by Texture
Map node to control particle generation.
Like I have one object with several clusters and each cluster have its own
uv and texture map. so how to generate particles filtered by different
texture maps from
Hello folks,
In March of last year, 2 colleagues and I left our jobs as software
developers in the games industry to form our own company - Redshift.
Our goal was to apply our experience with graphics hardware to the problem
of offline rendering.
Artists friends had been asking us for years why
congrats to you and your team! i was wondering when we would see/hear about
your work.
it would be great to see a video demonstration of redshift in softimage.
On Wed, Mar 13, 2013 at 11:12 AM, Nicolas Burtnyk nico...@redshift3d.comwrote:
Hello folks,
In March of last year, 2 colleagues and
Thanks Leonard. Very generous with your time and energies :-)
On Wed, Mar 13, 2013 at 6:02 AM, Cristobal Infante cgc...@gmail.com wrote:
Hi Leonard,
Thanks a lot for putting this tool together, I had a look at your 7
tutorial and it does look like a very handy tool.
In terms of how you
Thanks Steven!
We'll try to whip something together. Our team consists of 3 software
developers and we've been plugging away at the code side of things pretty
hardcore for the last year. We're admittedly very light on content,
something we hope to rectify during this alpha period - in some part
Couldn't you put the texturemaps on clusters of where you want them to
affect? Make a cluster of the side polys for the checkerboard one, and
another cluster for the text.
On Wed, Mar 13, 2013 at 1:48 PM, Debdas Mohanty devj...@gmail.com wrote:
I did that... But its Just adding every thing
The classroom is _really_ 2min render ?
Congrats to you, sending a request :)
Le 13/03/2013 19:18, Steven Caron a écrit :
congrats to you and your team! i was wondering when we would see/hear
about your work.
it would be great to see a video demonstration of redshift in softimage.
On Wed,
Yes - I'll try to make a video of that if I can get set up correctly for it.
Note that this is 2 mins on a GTX 470 which is nothing special in terms of
GPUs. You can expect significantly better times with a GTX 580 for
example. I don't have official times for that card, but I'd guess under
1.5
I did all that... maybe i am missing some thing...I am sending you an
example file pls check it...
Congrats on your achievement, looking forward to see more of it.
Just a quick question. For most of the render farm setups, usually the
machines do not have high-end GPU's on them for the obvious reason that
farm nodes never get user interaction and hence no need for realtime
graphic
That's looking quite interesting, Nicolas! Can't wait to know more!
-Tim C.
Magnetic Dreams
On 3/13/2013 1:33 PM, Nicolas Burtnyk wrote:
Yes - I'll try to make a video of that if I can get set up correctly
for it.
Note that this is 2 mins on a GTX 470 which is nothing special in
terms of
Exactly! This is VERY interesting. Hope I can test (sent request
already) :)
On 3/13/2013 12:33 PM, Nicolas Burtnyk wrote:
Yes - I'll try to make a video of that if I can get set up correctly
for it.
Note that this is 2 mins on a GTX 470 which is nothing special in
terms of GPUs. You can
Thanks Alok!
That'a an excellent question.
We don't currently have any official support for farm rendering. That
being said, with a suitable render manager or some hand rolled scripts, I
don't see why you couldn't have multiple machines working on different
frames of an animation. And of
Congrats on getting this up and going. I signed up for access. Do you
have any ballpark for where you think your price point will come in at?
On Wed, Mar 13, 2013 at 3:18 PM, Nicolas Burtnyk nico...@redshift3d.comwrote:
Thanks Alok!
That'a an excellent question.
We don't currently have
grats on progress.
I guess that renderer is tightly bind to CUDA and OpenCL versions like for
AMD cards are not in plan?
On Wed, Mar 13, 2013 at 8:23 PM, Daryl Dunlap twinsnakes...@gmail.comwrote:
Congrats on getting this up and going. I signed up for access. Do you
have any ballpark for
I would explode the filter compound – and make a chain of conditions, with
“if”, “and” and “or” nodes. (which can receive more than just two inputs by the
way)
also, on the clusters, there is an “isclusterelement” value – which should help.
From: Debdas Mohanty
Sent: Wednesday, March 13,
http://www.redshift3d.com/support/faqs/
:)
On Wed, Mar 13, 2013 at 12:32 PM, Mirko Jankovic
mirkoj.anima...@gmail.comwrote:
grats on progress.
I guess that renderer is tightly bind to CUDA and OpenCL versions like for
AMD cards are not in plan?
On Wed, Mar 13, 2013 at 8:23 PM, Daryl
Redshift does indeed use CUDA and is therefore currently limited to NVidia
hardware.
However OpenCL support is definitely in our plan, though not in the
immediate term (in other words it's not something we're working on at the
moment).
On Wed, Mar 13, 2013 at 12:32 PM, Mirko Jankovic
Steve, thanks for the link to the FAQ. I'm also pleasantly surprised to
see some Documentation there as well. :-]
On Wed, Mar 13, 2013 at 3:43 PM, Nicolas Burtnyk nico...@redshift3d.comwrote:
Redshift does indeed use CUDA and is therefore currently limited to NVidia
hardware.
However OpenCL
Looking forward to seeing some more samples, just wondering why you
continue to support 32-bit, given that autodesk have dropped support in the
more recent versions? Does it add much overhead? Scuse my uninformed
ponderings ;)
On 13 March 2013 19:43, Nicolas Burtnyk nico...@redshift3d.com
Yes we have some docs up on the site too! Thanks Daryl for mentioning that.
Please be aware though that the docs are not 100% complete and not super
clean. But they do contain lots of good info and should definitely help
people get started.
On Wed, Mar 13, 2013 at 12:55 PM, Daryl Dunlap
Great stuff! With the proper low/no GI configuration, would it be possible
to use Redshift to do near-realtime playback in Softimage?
Congratulations, too, for your achievements! It's so cool to see how
small teams of idealists can take the technological lead... wouldn't you
expect something like this from nVidia themselves?
Am 13.03.2013 21:03, schrieb François Painchaud:
Great stuff! With the proper low/no GI
We support 32 bit because it's not any more difficult to do so and doesn't
impact the performance or anything on 64 bit.
Basically, why not? :)
On Wed, Mar 13, 2013 at 1:00 PM, Matt Morris matt...@gmail.com wrote:
Looking forward to seeing some more samples, just wondering why you
continue
Ain't that some interesting piece of hardware:
http://www.digitaltrends.com/mobile/panasonic-20-inch-4k-windows-8-tablet-hands-on/
http://gadgets.ndtv.com/tablets/news/panasonic-shows-of-a-20-inch-4k-windows-8-tablet-at-ces-315261
Like a replacement for the HQ viewport? Ha!
Which rises the question: does it support progressive refinement?
From http://www.redshift3d.com/products/redshift :
1-Click Progressive Rendering
Redshift provides a progressive rendering mode that allows you to conveniently and
interactively
Nicolas, the docs say you support model instancing. But, do you also
support instancing via PointClouds (Particle instancing)?
On Wed, Mar 13, 2013 at 4:21 PM, Nicolas Burtnyk nico...@redshift3d.comwrote:
We support 32 bit because it's not any more difficult to do so and doesn't
impact the
Thanks Peter,
Got it...working fine with few IF, and ,or nodes,
Cls.polygon.isElement solved my problem...
Thanks Alan...
On Thu, Mar 14, 2013 at 1:05 AM, pete...@skynet.be wrote:
I would explode the filter compound – and make a chain of conditions,
with “if”, “and” and “or” nodes.
I want to create a Custom Parameter set with a pull down selection list
similar to what is available in Render Tree Property Pages with the Combo
lists. Is this possible?
--
Joey Ponthieux
LaRC Information Technology Enhanced Services (LITES)
Mymic Technical Services
NASA Langley Research
Cool screen, but the pen looks like a fat crayola!
Date: Wed, 13 Mar 2013 21:24:05 +0100
From: softim...@keyvis.at
To: softimage@listproc.autodesk.com
Subject: Panasonic 20 4K windows 8 tablet
Ain't that some interesting piece of hardware:
Enums (e.g. selection lists) are supported afaik, but GridData widgets are not.
I want to create a Custom Parameter set with a pull down selection list
similar to what is available in Render Tree Property Pages with the Combo
lists. Is this possible?
--
Joey Ponthieux
LaRC Information
@François: Unfortunately due to some fixed overhead on our side and things
like scene translation, Redshift will not render at 30 frames per second
even with all features turned down/off. That being said, depending on
resolution and settings, you could get renders on the order of 1-5 seconds
or
At this moment Redshift does not understand particle instancing. However
we can add support for this very easily.
On Wed, Mar 13, 2013 at 1:25 PM, Daryl Dunlap twinsnakes...@gmail.comwrote:
Nicolas, the docs say you support model instancing. But, do you also
support instancing via
Yeah, it very straightforward, you read the Particle 'Shape' attribute.
Then compare that value against the ICEShapeType enum, if the ICEShapeType
is Reference, then you have a Particle Instance of a object in the scene.
You can then take that value and get the actual Object by various means,
Yes, Combo boxes and GridWidgets are supported by making a custom property then
modifying the PPGlayout with your desired controls. If you want an interactive
PPGLayout where widgets respond to user interaction (enable/disable based on
button click, for example), then you'll need to define PPG
Hmmm.. Perhaps some optical data files don't work as well as others.
I tried a few other C3d files and they are working.
Sometimes applying the motion files hangs until I manually crash Softimage,
other times things are peachy keen.
Does anyone use c3d in Softimage through motor these days?
Adam
Hey guys,
Thanks for the great responses. I think I've responded to everyone who
sent an alpha request, but if you think I missed you, please shoot me an
email to remind me :)
Also, I wanted to share some render times as we ran some more tests this
afternoon comparing the GTX 470, GTX 670 and
77 secs for the Living room, that's impressive!!
---
Ahmidou Lyazidi
Director | TD | CG artist
http://vimeo.com/ahmidou/videos
2013/3/14 Nicolas Burtnyk nico...@redshift3d.com:
Hey guys,
Thanks for the great responses. I think I've responded to
Looks like the laggiest crap I've ever seen.
He had to tap stuff thrice or more before it'd pop up several seconds
later, he pinch zoomed and nothing happened (moved on silently like a
champ), lines appeared with several frames worth of lag, and I don't even
wanna think what the battery life
im loving my upgrade from a 24 to a 27, the extra res/ppi is lovely
if those 36 4k Eizo's were a reasonable price, i'd jump at one. ditch the
dual screen setup, go back to a single, large screen
On Thursday, March 14, 2013, Raffaele Fragapane wrote:
Looks like the laggiest crap I've ever
I like mine for about half the stuff I do, I find it miserable and painful
whenever software isn't up to scratch, and while Linux is ok, windows is
horrid with high res and the stupid 100/125/150% zoom options and its
crippled, for dummies, lack of interface customization (which most apps
ignore
I find the avoidance of GPU ram limitations to be quite something,
which is usually /a/ (if not *the*) major drawback of GPU rendering.
That along with aproximation methods which are also unusual,
made me quite curious.
Sent my trial request too :)
congrats!
Jason
On 13/03/2013 10:53 PM,
Hi
I'm trying to get values on geometry at a certain frame.
when using python - as long as I roll the playControl to that frame and
call refresh (ugly I know)
data= mesh.GetSurfacePointLocatorsFromPoints()
vertices=mesh.evaluatePositions(data)
I get the correct data.
on C++ no matter what I do,
51 matches
Mail list logo