Hi All,
There is no error when I trained my code on GPU and used the dtype=torch.float16
.
But when I run my code on CPU, I got the error
RuntimeError: “tanh” “_vml_cpu” not implemented for ‘Half’
I assume because torch.half
isn’t supported by torch.tanh
on CPU?
Is there any method can do tanh
in dtype=torch.float16
?
Thank you!