Due to compatibility issues, I am using pytorch=0.2.0 with python=2.7
I installed it using conda install pytorch=0.2.0 cuda80 -c soumith as it was pointed out on the forum that this will lead to reduction in lag while using .cuda() for the first time, however I do not see any improvements and loading still takes ~3mins.
Has been some time since PyTorch 0.2, but I can’t remember it being that slow. It should take a few seconds at most, not minutes. Not sure what’s causing the problem in your case, could be that it’s been a bug in PyTorch 0.2?
PyTorch 0.3 and 0.4 also work with Python 2.7, regardign the compatibility issues you mentioned, the Tesla V100 should work with cuda 8 & 9, so I think so if you can install it 0.3 or 0.4 somehow alongside the 0.2 version, it would help figuring out whether it’s a PyTorch 0.2-specific bug or sth else.
I remember this issue occurred if a wrong CUDA version was installed and in the first run it’s recompiling pytorch for your GPU.
Unfortunately there is no CUDA9 for pytorch 0.2.0.
However, could you print torch.version.cuda after your first cuda run?
It seems that I’m facing this issue using nvidia-docker2, though that maybe you can assist to figure it out.
Particularly, may it be that the nvidia-docker2 creates such recompiling because of its design ?
Can you take a look stackoverflow issue.
It’s not a pure pytorch issue, but maybe an nvidia-docker one, but maybe from a fast look at nvidia-docker design you could make some efficient conclusion for this issue in my setup and environment.
Particularly, do you think I should try install CUDA8.0 drivers on my host and try to see if I can make the connection between those drivers and the docker container (not sure if it is possible to manage in such way drivers, but that’s another question for CUDA I guess)?
Oh, forgot my Ubuntu 20 doesn’t support CUDA8.0.
Well, anyway it could be useful if you could take a look and tell me on which end in that case does seem to be the problem.