Hi,
Suppose we have a neural network, and we optimize the corresponding loss using Adam. Let
\Delta_w = w_{t+1} - w_{t}
I am interested in finding out
\Delta_w / w_{t}
Is there a way to do that?
Hi,
Suppose we have a neural network, and we optimize the corresponding loss using Adam. Let
\Delta_w = w_{t+1} - w_{t}
I am interested in finding out
\Delta_w / w_{t}
Is there a way to do that?