Daniel,

Using MPI and CUDA together is pretty much work in progress. On the
> developer branch, LinearAlgebra::distributed::Vector supports
> storing its elements on a CUDA device, but the values are copied to the
> CPU for MPI communication.
> The pull request https://github.com/dealii/dealii/pull/7303 allows to use
> CUDA-aware MPI instead but the memory transfer needs to be optimized.
>

If I could get involved in this development, I would be helpful to optimize
the memory transfer between GPU and CPU by testing different CUDA C
programming schemes like this:
https://devblogs.nvidia.com/how-overlap-data-transfers-cuda-cc/

In fact, as a principal investigator at Taiwan, I am qualified to submit
jobs into Taiwania, a recently established super computing machine
established in 2017, equipped with NVIDIA Tesla P100 SXM2. Moreover,
sometime next year, we will have another super computing machine available
to use (Taiwania 2), which now ranks 20th in the world with NVIDIA Tesla
V100 SXM2.

https://www.top500.org/system/179590

Sincerely,

Chih-Che

-- 
The deal.II project is located at http://www.dealii.org/
For mailing list/forum options, see 
https://groups.google.com/d/forum/dealii?hl=en
--- 
You received this message because you are subscribed to the Google Groups 
"deal.II User Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to dealii+unsubscr...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to