I’m using adam for optimization. Should I change learning rate using this?
for param_group in optimizer.param_groups: param_group['lr'] = lr
It seems every time I change the learning rate, the loss increases a lot, and the accuracy goes down at the learning rate transition point. What’s the reason?