Does Thrust support cuda-aware mpi for multi gpu executions? #1039
Unanswered
manasi-t24
asked this question in
Thrust
Replies: 0 comments 1 reply
-
|
Thrust algorithms do not yet support using multi-GPUs, and likewise they do not use MPI or any communication library underneath. The only data transfers in Thrust would be host <-> device transfers for a single GPU. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I am going through Nvidia blogs regarding optimization of GPU code.
I came across this post https://devblogs.nvidia.com/introduction-cuda-aware-mpi/ about using cuda-aware mpi to speed up the data transfer between two GPU's (either on the same node or on different nodes).
I have an application RapidCFD which uses thrust libraries for multi GPU solving of a system of linear equations. When I profiled the application, I saw that a lot of time is being consumed in data transfers.
Since, the data transfer details are hidden to the user because thrust library is used, I want to know if thrust employs all the data transfer optimizations mentioned by Nvidia in their blog posts, like the above one about direct data transfer between 2 GPU's using cuda aware mpi and this one (https://devblogs.nvidia.com/how-optimize-data-transfers-cuda-cc/) about allocating storage in the pinned memory.
Any help would be appreciated.
Regards,
Manasi
Beta Was this translation helpful? Give feedback.
All reactions