I am training a self-supervised learning model. The
training loss is moving within a range of values, even after training for 50+ epochs. And not decreasing after a certain point.
I want the model to overfit. I have set dropout to 0. Will try playing with the learning rate. What else can be done?
Training loss currently