I am following the tutorial from oficial pytorch website, where teaches how to fine tunning with mask-rcnn
I have a question in the Main function (which is little down below).
You will see this code
params = [p for p in model.parameters() if p.requires_grad]
As far i know, when we need do fine tuning classification models (like VG 16), we freeze every layer except the last one. Why here we use this code?
I have already readed the official paper of the faster-rcnn where it described until 3 different ways to train this architecture
Please i will appreciate any explication
PDT: this is my first project with this kind of architecture
This line of code stores all parameters, which require gradients and will thus be trained, into a list.
This list is then passed to the optimizer. You could also pass all parameters to the optimizer and let the check internally, if valid gradients were calculated for each parameter or if this parameter should be skipped.
The posted code snippet is a good way to write explicit code in my opinion.
Thanks for your answer, i appreciate i too much but my doubt hasnt been solved yet
After of reading the oficial Faster RCNN paper
The documents describe 3 different methods
The method that did you explained me is one of the training method explained in the paper?
If not, why Pytorch is using this method of training?
Sorry whether i am asking something trivial but now i am in a point where i beginning to understand lot of things (and questions come)