Considerable absolute error in torch.matmul

Hi Koramajin!

I made a mistake in my earlier post – see the edit, above.

Please try specifically setting the TF32 flag to False:

torch.backends.cuda.matmul.allow_tf32 = False

and see if that resolves the issue.

(Sorry for the earlier mistake.)

Please note: My understanding is that your A100 gpu does support TF32,
so this behavior would be expected.

Best.

K. Frank