ReduceLROnPlateau reduces LR without need


I have a problem with ReduceLROnPlateau. I initiated it using:
scheduler = torch.optim.lr_scheduler.ReduceLROnPlateau(optim, verbose=True, patience=5)

and then, after each epoch (validation), I perform

The problem is, that the scheduler decreases LR even though the loss does not plateau.


Can somebody please help me out?

Thank you very much! :blush:

Could you print the values of loss_meter_validation.mean for the first steps, which seem to trigger the decrease in the learning rate, please?