I know it’s very difficult to debug without context, but I’m asking just in case this is a known phenomenon. I’m training a U-Net architecture for regression. During training, I frequently see discontinuities in the loss funciton, which look very strange to me. Here’s how the training history looks like:
I’m using Adam optimizer with a learning rate of 1e-3.