Autograd Why accumulation gradient mechanism?

Hi,

I’m not sure to understand your question.
If you ask why .backward() accumulates gradients, then this post should give you a good idea: Why do we need to set the gradients manually to zero in pytorch?