Do we need to zero_grad
the network too? As net
isn’t wrapped by the optimizer
.
optimizer = torch.optim.Adam([x], lr=1)
net.eval()
.
.
optimizer.zero_grad()
net.zero_grad()
loss.backward()
optimizer.step()
according to below link:
Do we need to zero_grad
the network too? As net
isn’t wrapped by the optimizer
.
optimizer = torch.optim.Adam([x], lr=1)
net.eval()
.
.
optimizer.zero_grad()
net.zero_grad()
loss.backward()
optimizer.step()
according to below link: