Did you run your cuda settings in one line like this:
CUDA_VISIBLE_DEVICES=1 python myscript.py
Could it be that you typed it before calling your python script, so that it has no effect and your models are still pushed on GPU 2, 3 and 4. This would explain the same memory amount of these GPUs.
However, if you really want to use only GPUs 1, 2 and 3, place the setting in one line and use
Could you check in which line the error was thrown and paste this code snippet?
You have to make sure, that all operations are executed on the same GPU id.
For example if your generator and discriminator are placed on different GPUs, you have to push the Variables to the according GPU:
Im running in a notebook… so there’s no notion of running the script like that. I see what the issue might be there. But I think I handle that. What’s happening for me when I use DataParallel is that the first N-1 GPUs have memory allocated to them, but the program keeps waiting for the Nth GPU to have the requisite memory allocated… Not sure if thats a common issue?