I have a trained CNN with 70% accuracy, I apply transfer learning, add Global Average Pooling to it, and finetune this new model. However, the train & validation accuracy differ wildly depending on PyTorch’s random weight initialisation values (I’ve seen from 10% to 50% so far), even across different learning rates (
1e-06). The train & valid loss are simply stuck at around their initial value, even after 400 epochs of backpropagation.
I have made sure to set the model’s parameters with
requires_grad() = True. What could be a possible reason for this bizarre behaviour? Thank you!