I’ve been used Caffe which has the functionality of change the learning rate call ‘lr_policy’. This functionality is very convenient. I want to know whether there is a similar way like lr_policy’’ in pytorch Or should I code it by myself?
Yes, there is a group of function called lr_schedular. Please refer to this page: http://pytorch.org/docs/0.3.0/optim.html#how-to-adjust-learning-rate
Thank you for your help!