Hi,
While training my model I got NaNs as the result of loss function after many iterations. I used anaomly_detection
to see what is causing that issue and I got that error.
Function 'LogBackward' returned nan values in its 0th output
This is the only log funtion I used in my forward path
return o.log()
And that is a sample of the values it receives
[0.0004, 0.0006, 0.0010, ..., 0.0143, 0.0143, 0.0143],
...,
[0.0007, 0.0005, 0.0007, ..., 0.0146, 0.0146, 0.0146],
[0.0007, 0.0005, 0.0007, ..., 0.0147, 0.0147, 0.0147],
[0.0007, 0.0005, 0.0007, ..., 0.0147, 0.0147, 0.0147]],
I cannot catch any zeros or negative numbers passed to it. Can you maybe tell me if anyline I should use to catch the number that is causing that issue or anyways to recover from this?