Backprop in PyTorch?

Do we need to zero_grad the network too? As net isn’t wrapped by the optimizer.

optimizer = torch.optim.Adam([x], lr=1)
net.eval()
.
.
optimizer.zero_grad()
net.zero_grad()
loss.backward()
optimizer.step()

according to below link: