The following is from official tutorial. My question is: what’s the point of no_grad here? My understanding is that it is only useful if you want to save costs while in running the forward pass.
with torch.no_grad():
for param in model.parameters():
param -= learning_rate * param.grad