Tesla T4 with CUDA capability sm_75 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_70.
How come the pre-built PyTorch binaries for 1.6 do not support the newest CUDA capability? Is this intentional? How can I fix my setup? Thanks!
I’m able to use but not sure whether all CUDA functionalities work as intended because of this error. This is a strange behavior. Building Pytorch isn’t easy thing on my restricted server.
Facing the same issue here. Card is GeForce RTX 2080 Ti. What does this warning mean? Inside torch, torch.cuda.is_available() returns True and I can train with GPU. Do I need to and how can I fix it?
Thanks in advance.
/home/zbxs/miniconda3/lib/python3.7/site-packages/torch/cuda/__init__.py:125: UserWarning:
GeForce RTX 2080 Ti with CUDA capability sm_75 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_37 sm_50 sm_60 sm_70.
If you want to use the GeForce RTX 2080 Ti GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
warnings.warn(incompatible_device_warn.format(device_name, capability, " ".join(arch_list), device_name))
What is the rationale behind removing sm_75 for the binaries that target CUDA 10.1? I am afraid that I cannot upgrade to CUDA 10.2 due to dependency and driver issues.
What are the effects of using the PyTorch binaries that do not support sm_75 on a graphics card that is supporting sm_75?
Just a heads-up for everyone following along, the issue has been fixed underneath the hood but the binaries have unfortunately not been update yet: https://anaconda.org/pytorch/pytorch/files
The new binaries seem to be available now, at least for pip. Make sure to use --no-cache-dir to not use a locally cached binary when pip installing PyTorch again.
The sm_70 argument should add the sass and ptx to the binaries. The ptx would make the code executable for your T4 (compute capability 7.5), but might not yield the optimal performance.
The Turing architecture is based on Volta’s Instruction Set Architecture ISA 7.0, extending it with new instructions. As a consequence, any binary that runs on Volta will be able to run on Turing (forward compatibility), but a Turing binary will not be able to run on Volta. Please note that Volta kernels using more than 64KB of shared memory (via the explicit opt-in, see CUDA C++ Programming Guide) will not be able to launch on Turing, as they would exceed Turing’s shared memory capacity.