When the training is progressing, I see the test accuracy does not change a lot and is almost steady. However, when I save the checkpoint and resume the iteration from the checkpoint the training show a jump in the test accuracy with a significant margin.
I wonder if you guys know what is going on and possibly provide an explanation for such a random behavior. @ptrblck
It’s hard to understand the situation in detail without any code; are you referring to test accuracy that is run on a separate test dataset?
It might be useful to verify that the test accuracy is consistent when saving a checkpoint vs. when loading the same checkpoint (without any additional training).