Hi,
I’m not sure to understand your question.
If you ask why .backward()
accumulates gradients, then this post should give you a good idea: Why do we need to set the gradients manually to zero in pytorch?
Hi,
I’m not sure to understand your question.
If you ask why .backward()
accumulates gradients, then this post should give you a good idea: Why do we need to set the gradients manually to zero in pytorch?