You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Shared memory MPI instances is not mutually exclusive from each MPI rank acknowledging a single (unique) device.
Current workaround is to ensure proper device mapping prior to integrator call based on known MPI <-> device affinity (i.e. replicate this according to a known affinity)
The text was updated successfully, but these errors were encountered:
wavefunction91
changed the title
Need a robust check for MPI <-> device mapping in CUDA
[CUDA] Need a robust check for MPI <-> device mapping in CUDA
Oct 2, 2020
The check introduced in 7ba2f43 and reverted in 94a1f86 is not robust. Fails for the following resource configuration on Summit https://jsrunvisualizer.olcf.ornl.gov/?s1f0o11n6c7g1r11d1b27l0=
Shared memory MPI instances is not mutually exclusive from each MPI rank acknowledging a single (unique) device.
Current workaround is to ensure proper device mapping prior to integrator call based on known MPI <-> device affinity (i.e. replicate this according to a known affinity)
The text was updated successfully, but these errors were encountered: