Why when I redefine my model to have fewer parameters and retrain (after training once before) does PyTorch sometimes through an OOM error?
In other words: How do I tune model hyperparameters without running into OOM errors?
Do I have to reset the computational graph / saved gradients in the dataset’s tensors? If so, how?