Cool. We have quite a few computers and probably can share some wisdom
Has anyone ever actually tested Cycles with 9 GPUs simultaneously? We’re using GTX 1060s, 6GB. Do they need to be full speed PCIE slots, or does 4x work (we’re using gen 2 risers to fit all the GPUs on one board)? I’d imagine it depends on the complexity of the scene.
So the way cycles works is that multiple gpus cannot work on the same tile at once, and gpus love large tile sizes. So… the best thing to do is if you have multiple gpus in the computer at the same time, is to run multiple blenders rendering on single gpus each. Python scripting is the easiest way of getting this setup. Our results were with 2 1080’s below:
1x 1080 = 100%
2x1080, rendering same frame = 140%
2x1080, rendering different frames in different blenders = 189%
You also mention whether they need to be full speed pci-e slots. Right now, half of our render grid are using 1 x PCI-E 1 speeds, and it impacts the speed over a full 16x PCI-E 2 slot by about 1-2 seconds. So no doesnt make a huge difference
We’d also like to know if there are any scripts for remotely controlling the rendering, so that we don’t manually have to set the master and slave every time we do a render. The rigs are running Ubuntu. In a perfect world, I’d be able to hit one button and 1. stop the mining, 2. boot to the blender OS, and 3. start blender to begin rendering over the network
Hmm seems like you are using netrender to render? We tried it out back in the day and couldnt get it to efficiently work. We tend to use the stock standard placeholder/no overwrite option all rendering to the same location.
The way we set it going over 18 computers, is a windows batch script to log onto each computer and run a shell script (you could probably do this via ssh aswell)… which then does what you want, in your case, kills any cryptomining and starts it in slave mode. I am 90% confident there is a python command to run as soon as it loads up.
After running the BMW benchmark, we found that our rendertime was 1 minute 10 seconds.
Keep in mind, the system has 8xGTX1060s, so while fast, the test was slightly disappointing. I see on blenchmark that the top speed is 13 seconds, and that’s with 5xGTX980s. So why am I seeing such slow speeds? My only guess is that the gen 2 risers are the bottleneck, but even then it only takes 10 seconds to synchronize. Anyone have any thoughts on this?
Blenchmark is not a good measurement for benchmarking.