I have three GPU’s and have been trying to set CUDA_VISIBLE_DEVICES
in my environment, but am confused by the difference in the ordering of the gpu’s in nvidia-smi
and torch.cuda.get_device_name
Here is the output of both:
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 387.34 Driver Version: 387.34 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
|===============================+======================+======================|
| 0 TITAN Xp Off | 00000000:02:00.0 On | N/A |
| 23% 38C P8 17W / 250W | 811MiB / 12188MiB | 0% Default |
+-------------------------------+----------------------+----------------------+
| 1 Graphics Device Off | 00000000:03:00.0 Off | N/A |
| 34% 49C P8 26W / 250W | 0MiB / 12058MiB | 0% Default |
+-------------------------------+----------------------+----------------------+
| 2 Graphics Device Off | 00000000:04:00.0 Off | N/A |
| 28% 40C P8 24W / 250W | 0MiB / 12058MiB | 0% Default |
+-------------------------------+----------------------+----------------------+
>>> torch.cuda.get_device_name(0)
'Graphics Device'
>>> torch.cuda.get_device_name(1)
'TITAN Xp'
>>> torch.cuda.get_device_name(2)
'Graphics Device'
I would have expected the device numbers to be consistent across these applications. If not, then what should I expect if I am using another library like keras or tensor-flow?
Thanks in advance.