You don't need anything bigger as a matter of fact I'm running that exact
PSU for 4x Titan X, and before them it was pushing 4x Titans.
Nothing more is needed. 1200W is cutting it a bit too close but 1500 is
perfectly enough.

On Thu, Aug 6, 2015 at 1:31 PM, Morten Bartholdy <x...@colorshopvfx.dk>
wrote:

> Thanks guys for the very comprehensive walkthrough on this topic - this
> list is still a goldmine! :)
>
>
>
> So for a 4 GPU workstation setup - would you suggest something like the
> Corsair AX1500i or should it be bigger?
>
>
>
>
>
> Morten
>
>
>
>
> Den 6. august 2015 kl. 12:49 skrev Matt Morris <matt...@gmail.com>:
>
> The conversation was aimed towards renderfarms rather than workstations
> though, and I imagine running a render job per gpu rather than per node, so
> that the scaling per gpu is much better (ie 100% minus maybe a small hit on
> the cpu usage being shared). Could be run headless so no need for a display
> card.
>
> In terms of power at the wall, in the uk a kettle will routinely use 3000w
> (albeit only for a short time) so a 4 gpu pc should be within acceptable
> limits - between 1000 - 1500 w when rendering. The biggest problem I've had
> is finding a suitable UPS which is silent as most at that rating need fans,
> and are designed to sit in a server room instead of a studio space.
>
> There was an interesting post on the RS forums recently from a guy setting
> up a gpu renderfarm using these:
> http://www.supermicro.com/products/system/2U/2028/SYS-2028GR-TRH.cfm
>
> dual xeon, 6 gpu solutions mmm. Sounds like quite a bit of work to get it
> all working smoothly though, including modifying 980ti card power outlet
> from top to back to match tesla cards.
>
>
> On 6 August 2015 at 10:16, Tim Leydecker < bauero...@gmx.de > wrote:
>
> Would you guys find the 980Ti hitting the sweetspot between price and
> performance?
>
> How about connectors and power supply?
>
> The 970 is running on 2x6pin, e.g. a maximum of 150 Watts plus the 75
> Watts from the slot, a 225 Watts total.
>
> The 980ti is mostly 1x6pin and 1x8pin, the 1x8pin offering 150Watts
> compared to a 1x6pin offering 75 Watts.
>
> In my case, I find it already hard to provide more than one 1x8pin and
> 1x6pin via connectors.
> How do you guys provide reliable power to more than 1 or 2 graphics cards
> without melting your power lines?
>
> Here in Germany, it is rare to have more than around 1 kW sustained drain
> per average wall plug supported by a great many home installations.
> There is always loads of headroom of course but technically, constantly
> draining a lot more from  such a wall plug can get, uhmmm, hot.
>
> That´s a few of the reasons I suggested to start out with just 1 card,
> like a Titan X (or a GTX980ti), case power supply connection, wall plugs,
> electrical limits.
>
> Cheers,
>
> tim
>
>
>
>
>
>
>
>
> Am 05.08.2015 um 16:10 schrieb Mirko Jankovic:
>
> agree. 980ti is just a bit above 2 970s price wise, performance wise it
> realyl dpends on scenes you are working on. but I plan to upgrade my 4x970
> with 980ti as soon as possible, even if it means replacing 1 by  1
>
> On Wed, Aug 5, 2015 at 3:36 PM, Matt Morris < matt...@gmail.com > wrote:
>
> The 970 is the most cost efficient only with scenes that fit into its
> memory - which using redshift is limited to 3.5Gb because of the internal
> memory architecture. I'd recommend looking at gpus with 6Gb or higher. The
> 980ti is a great card for the money, and the extra vram will help
> performance even on small scenes as you can utilise memory optimisation
> settings. Because you're limited to 4 gpus (risers don't work too well and
> limited by number and speed of pci-e lanes as mirko said) you want to make
> the most of that space. Per card electricity usage and heat output isn't
> that much more for the 980ti.
>
> On 5 August 2015 at 14:04, Tim Leydecker < bauero...@gmx.de > wrote:
>
> Thanks for the clarification, Dan.
>
> I think I mixed this up with the download section of the forum for
> customers?
>
> Whatever, good that the registered user forum is accessible to interested
> parties.
>
> Cheers,
>
> tim
>
> P.S: For Hair, Shave&Haircut is supported (I don´t have personal
> experience with it).
>
>
> Am 05.08.2015 um 14:17 schrieb Dan Yargici:
>
> " you may find it helpful to register in the Redshift3D.com forums, afaik
> you´ll need to have
> at least one registered license to get access to the "Registered users
> only" forum area."
>
> Just to clear this up.  I'm pretty sure you don't need to have a license
> to access the Registered Users section of the Redshift forums.
>
> DAN
>
>
> On Wed, Aug 5, 2015 at 2:58 PM, Rob Chapman < tekano....@gmail.com > wrote:
>
>
> A lot of good and informed points by all, just wanted to add, this guy
> here, Sven, at http://www.render4you.de/renderfarm.html recently became
> the first official Redshift GPU render farm and have used him already on a
> few jobs with very tight deadlines.  Essentially he has a rack of 7x Tesla
> K40st - so 1 node is the equivalent of a 6x single 980gtx which I find is
> pretty cost effective solution of adding a decent online GPU render node,
> that works with hardly any setup if you have a redshift scene ready to go
>
>
> best
>
> Rob
>
> On 5 August 2015 at 11:56, Tim Leydecker < bauero...@gmx.de > wrote:
>
> Hi Morten,
>
> you may find it helpful to register in the Redshift3D.com forums, afaik
> you´ll need to have
> at least one registered license to get access to the "Registered users
> only" forum area.
>
> There´s a few threads there about Hardware, multiple GPU systems and some
> user cases
> of testing single gpu vs. multi gpu rendering plus some Developer info
> about roadmaps and such.
>
> Personally, I´m a big fan of Redshift 3D.
>
> Still, here´s a few things to consider you may find useful:
>
> - Compared to Arnold, there is no HtoA or C4DtoA equivalent, e.g. no
> direct C4D or Houdini support
> - Compared to Arnold, rendering Yeti is not yet supported in Redshift3D -
> it´s looked at, no ETA.
> - Maya Fluids, Volumerendering, FumeFX e.g. Fire&Smoke&Dust&such isn´t in
> Redshift3D sofar
>
> - Multitasking, compared to CPU based multitasking and task switching
> (e.g. switching between
>   rendering in Maya, Softimage while simultaneously comping in Nuke and
> painting Textures in Photoshop
>   or Mari) may pose GPU specific limitations with multiple applications
> fighting for a very limited GPU VRAM.
>  Redshift3D can utilize system RAM for VRAM but there can be headache when
> other, "dumber" apps go ahead
>  and just block VRAM for their caching. It´s well worth running a good few
> hard tests in typical workflow scenarios.
>  Maya, Substance Painter/Designer, Nuke, Photoshop, they all offer one
> type or another of GPU caching or GPU
>  acceleration option. My personal feeling is, such stuff never gets tested
> in real-world, multiple-applications-running scenarios.
>
> At a glance, it would sound easy enough to have separate, dedicated GPUs
> run headless for rendering and reserving one GPU
> for viewport display and other apps but to be honest, all this stuff is so
> new, even thought it´s great, it´s still pushing grown
> legacy workflows and boundaries and in doing so, it may sometimes hurt.
>
> My very personal suggestion is:
>
> - a starter kit is just one GPU, optimally a Titan X with 12GB VRAM.
> - step 2, adding a second GPU, running headless, reserved for rendering
> - step 3, adding a third GPU, comparing speed to step 2
> - step 4, price/performance balancing, comparing a 1-2-3 GPU GTX970 render
> rig with the above
>
> Could be you find out you like to run 1 Titan X for viewport display and
> multi-apps, and 2 GTX970 for a render job.
>
>
> Another thing.
>
> Multi-socket CPU boards and PCIe slots. It seems easier to get solid
> single socket CPU boards with lot´s of PCIe slots.
>
> Again, from my personal experience running a current generation dual
> socket Xeon rig, it is annoying how many CPU
> cycles I see wasted away in idle in most of my daily chores, except for
> pure rendering with Arnold or the likes, I find
> myself mostly having one CPU and even most of the other CPU´s cores just
> not used properly by software.
>
> I think a good sweetspot would have been to just go for one fast, solid
> 6-core(budget) or 8core (current) CPU, unless of course for a dedicated
> render slave...
>
>
> Cheers,
>
> tim
>
>
>
>
>
>
>
>
>
>
>
> Am 05.08.2015 um 12:05 schrieb Morten Bartholdy:
>
> I know several of you are using Redshift extensively or only now. We are
> looking in to expanding our permanent render license pool and are
> considering the pros and cons of Arnold, Vray and Redshift. I believe
> Redshift will provide the most bang for the buck, but at a cost of some
> production functionality we are used to with Arnold and Vray. Also, it will
> likely require an initial investment in new hardware as Redshift will not
> run on our Pizzabox render units, so that cost has to be counted in as well.
>
>
>
> It looks like the most priceefficient Redshift setup would be to make a
> few machines with as many GPUs in them as physically possible, but how have
> you guys set up your Redshift renderfarms?
>
>
> I am thinking a large cabinet with a huge PSU, lots of cooling, as much
> memory as possible on the motherboard and perhaps 8 GPUs in each. GTX 970
> is probably the most power per pricepoint while Titans would make sense if
> more memory for rendering is required.
>
>
> Any thoughts and pointers will be much appreciated.
>
>
>
> Morten
>
>
>
>
>
>
> --
> www.matinai.com
>
>
>
>
> --
> www.matinai.com
>
>
>
>

Reply via email to