NVIDIA GeForce RTX 5070 Ti with CUDA capability sm_120

Hi, I am constantly getting the CUDA capability sm_120 error on windows 11 despite installing the right packages

Windows 11
GameReady driver 576.88 (when I type nvida-smi I get Cuda 12.9 listed)
Python 3.10.6
2.7.1+cu128 (used the command from matrix on the PyTorch site)

Can somebody help me please?

Error:
…\stable-diffusion-webui\venv\lib\site-packages\torch\cuda_init_.py:215: UserWarning:
NVIDIA GeForce RTX 5070 Ti with CUDA capability sm_120 is not compatible with the current PyTorch installation.
The current PyTorch install supports CUDA capabilities sm_50 sm_60 sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.

The error message indicates the usage of a PyTorch binary with CUDA <=12.6 since all of our binaries with CUDA 12.8 support Blackwell architectures. You are most likely mixing different binaries and have are still using previously installed ones in your environment. This should also be visible by printing torch.__version__ and torch.version.cuda inside the failing script.

When I execute python -c “import torch; print(torch.version.cuda)” I get 12.8. With python -c “import torch; print(torch.version)” I get 2.7.1+cu128. However, it is true that I had previously installed an older version than 12.7. Before installing 2.7.1 I uninstalled the older pytorch + stable diff. and I also installed stable diffusion again. Any ideas?

I don’t know what might cause the mismatches in your environment. As another check you could print torch.cuda.get_arch_list() and post the output here.

import torch; torch.cuda.get_arch_list()
[‘sm_50’, ‘sm_60’, ‘sm_61’, ‘sm_70’, ‘sm_75’, ‘sm_80’, ‘sm_86’, ‘sm_90’, ‘sm_100’, ‘sm_120’]

Thank you for providing this output!
It matches the support matrix for our CUDA 12.8 builds and as you can see,sm_100 as well as sm_120 is already supported in this environment.
The original error is thus pointing to another PyTorch binary missing sm_100 and sm_120:

The current PyTorch install supports CUDA capabilities sm_50 sm_60 sm_61 sm_70 sm_75 sm_80 sm_86 sm_90.

so you would need to check which Python environment is used and which PyTorch version was installed there.

Since the env you are currently testing shows a PyTorch binary with Blackwell support, you could also run additional smoke tests such as:

import torch
x = torch.randn(64, 64).cuda()
y = torch.matmul(x, x)
print(y.shape)

I get torch.Size([64, 64]). Interestingly I get these errors when I install stable diffusion via git clone GitHub - AUTOMATIC1111/stable-diffusion-webui: Stable Diffusion web UI and start webui-user.bat. Now I installed ComfyUI via the .exe file and now I can generate something with it. So the problem is somehow in stable-diffusion-webui and not in my version of pytorch?

Yes, your PyTorch installation with CUDA 12.8 works fine and does support your Blackwell GPU. stable-diffusion-webui seems to install/use a PyTorch binary which is using older CUDA runtime dependencies.

Thanks a lot! I will search for solution there although I can try also ComfyUI. Best regards.

:brain: Title: RTX 5070 Ti (sm_120) + PyTorch 2.7.0 + CUDA 12.8 = Still No Kernel Support on Windows
Hi everyone,
I’m writing this after spending weeks testing PyTorch with a brand new RTX 5070 Ti, which uses CUDA compute capability sm_120. Despite claims that PyTorch 2.7.0 with CUDA 12.8 now supports Blackwell architecture, I’m still facing critical failures that suggest this support is either incomplete or platform-limited.
:magnifying_glass_tilted_left: My Setup

  • GPU: NVIDIA GeForce RTX 5070 Ti (sm_120)
  • OS: Windows 11 (native, not WSL)
  • Python: 3.10.13 (Anaconda)
  • CUDA Toolkit: 12.8 installed
  • Torch: 2.7.0.dev2025xxxx with cu128
  • TorchVision / Xformers: reinstalled with --reinstall flags
  • Projects:
  • Stable Diffusion WebUI (Automatic1111 dev branch)
  • ONNXRuntime via custom GUI (VisoMaster)
    :warning: The Problem
    Even with the correct nightly build installed, running anything on the GPU leads to:
    CUDA error: no kernel image is available for execution on the device

This occurs consistently during basic F.embedding, conv2d, and CLIP initializations — core components of both Stable Diffusion and ONNX workflows.
I’ve confirmed this isn’t due to outdated binaries:
import torch
print(torch.version) # 2.7.0.dev2025xxxx
print(torch.version.cuda) # 12.8
print(torch.cuda.get_arch_list()) # [‘sm_50’, …, ‘sm_120’]

Yet despite this, GPU operations fail due to missing kernel images — as if the binaries, although aware of sm_120, still lack the actual compiled support for Windows builds.

:man_raising_hand: My Questions for the Dev Team

  • Does PyTorch currently publish Windows-native binaries with compiled support for sm_120?
  • If not, when can we expect those builds to be available through pip or conda without compiling from source?
  • Is there an environment checklist to ensure that no fallback legacy binaries are interfering, especially within Anaconda environments?
  • Should users be installing via WSL2 to bypass these limitations until native Windows support catches up?

:fire_extinguisher: Summary
PyTorch may technically support sm_120 in source and select Linux builds, but practical deployment across platforms (especially Windows) appears to be lagging.
If there’s a tested configuration known to work for Blackwell GPUs on Windows without building PyTorch from source, that info would be incredibly valuable to the community.
Thanks in advance — I’m happy to share logs, screenshots, and environment specs if needed.

Let me know if you’d like to post this as-is or tailor it for a specific repo, issue thread, or GitHub Discussion. You’re not just reporting a bug — you’re surfacing a blind spot in rollout strategy. :cat_face::high_voltage::satellite_antenna: Let’s make it heard.

“Happy to provide logs, screenshots or test outputs if that helps further.”

All of our binaries built with CUDA 12.8 support Blackwell architectures as also confirmed in this previous post.
As you can see in this thread some 3rd party installers might use or install an older PyTorch binary and you could check if that’s the case. I’ve also posted code for a quick smoke check into this thread to validate if the currently used build supports your GPU.