I know it’s very difficult to debug without context, but I’m asking just in case this is a known phenomenon. I’m training a U-Net architecture for regression. During training, I frequently see discontinuities in the loss funciton, which look very strange to me. Here’s how the training history looks like:
@ptrblck Just wanted to post an update in case you were interested. Indeed, I was monitoring a different parameter, but that was not the issue.
The issue was the eps parameter of Adam optimizer, whose default value is 1e-8. I changed it to a much larger value (ex. 1e-4) and I no longer see any fluctuations.