edmBernard commented on issue #7350: Multi-Training-Task on the same GPU card URL: https://github.com/apache/incubator-mxnet/issues/7350#issuecomment-320575891 You have two training at 27 samples/sec each one ? GPU have power processing so if one training use 100% the second need to wait even if they share the memory. ``` +-----------------------------------------------------------------------------+ | NVIDIA-SMI 375.66 Driver Version: 375.66 | |-------------------------------+----------------------+----------------------+ | GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. | |===============================+======================+======================| | 0 GeForce GTX 1070 Off | 0000:01:00.0 Off | N/A | | 29% 44C P8 14W / 151W | 0MiB / 8113MiB | 0% Default | +-------------------------------+----------------------+----------------------+ ^ | | ``` ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org
With regards, Apache Git Services