Transfer Learning train two last layer (Bloc)

it is possible to set params in layer before fully-connected layer to train params like case 2 in the table ?
thanks :slight_smile:

Yes, you can freeze specific layers and train others by manipulating the .requires_grad attribute of each parameter (setting it to False would freeze the parameter). The Finetuning tutorial would give you an example.