Hello everyone!
It is said that bfloat16 is only supported on GPUs with compute capability of at least 8.0, which means nvidia V100 should not support bfloat16.
But I have test the code below on a V100 machine and run successfully.
a = torch.randn(3,3,dtype=torch.bfloat16,device="cuda")
b = torch.randn(3,3,dtype=torch.bfloat16,device="cuda")
c = torch.matmul(a,b)
print(c.dtype)
print(c.device)
and get the result
torch.bfloat16
cuda:0
but when I run print(torch.cuda.is_bf16_supported()), I got False
So what is situation here?