To reduce the size of the precompiled binaries (that were going out of hand), we had to remove some of the old GPU architecture that are used by a minority of people.
You can still use it by compiling from source and it will still work as before !
We’ve got PyTorch CUDA bundles with compute capabilities starting 3.0, 3.5
There is no separate python package, but you can extract the package from the installer archive.
Since compiling from source is a bit of a headache, and I have a GPU with
a cuda capability of 5.0:
Does that message mean that Pytorch doesn’t support my GPU from 0.3.1 (which is the first version which print this warning AFAIK),or that it won’t support it going forward? (which is what deprecation usually mean)
Also,what kind of operations should I expect NOT to work?
Any reasonable way for me to tell when an unsupported operation was executed?
From 0.3.1 onward, cuda capability 5.0 will not be included in the pre-packaged binary release (so all torch.cuda related stuff will not work).
You will be able to get pytorch to work with such architecture by compiling from source (so all operations will work).
Yes that would be 0.3.0 but that means that you will be missing a lot of bugfix.
Compiling from source should be pretty straighforward if if you already have cuda installed. Let me know if you need help with that.
Compiling current master will give you as of writing what is going to become 0.4 in the future.
If you want to keep to stable release, you can git checkout 0.3.1 to get the exact state corresponding to the 0.3.1 release before compiling.
Yes it does give me a warning , but my code which runs mostly on a gpu works (unlike the 0.3.1 binary). Should I ignore the warning?
So there’s a better chance of the main branch having more bugs compared to the 0.3.1 branch?
If so I’ll recompile and reinstall it. Might be the last message of this thread so: Thanks a lot man, I really appreciate your help!
BTW you don’t have to install it from source.
You may simply do pip install --user http://download.pytorch.org/whl/cu80/torch-0.3.0-cp27-cp27mu-linux_x86_64.whl
and save yourself some time!
Torch-0.3.0 still supports my old GPU
Hi,
I’m a bit stuck with a M1000m gpu here. Current version is not supporting it any more:
Found GPU0 Quadro M1000M which is of cuda capability 5.0.
PyTorch no longer supports this GPU because it is too old.
Tried to compile from source (following the procedure from readme>installation>from source) but it didn’t solve it. I’m ok to use older binaries but the one suggested by @sangeet is not compatible… any other ideas?
ldd /home/../anaconda3/lib/python3.6/site-packages/torch/_C.cpython-36m-x86_64-linux-gnu.so
[....]
libmkl_gf_lp64.so => not found
libmkl_gnu_thread.so => not found
libmkl_core.so => not found
[....]
We are having a problem where we have a number of people on our team who are running NVIDIA Quadro M1200 cards, and the PyTorch error says that the card has CUDA capability 5.0, but the official NVIDIA page clearly says that the compute capability of this card is 5.2: https://developer.nvidia.com/cuda-gpus