Find exact step size in Adam

Hi,

Suppose we have a neural network, and we optimize the corresponding loss using Adam. Let

\Delta_w = w_{t+1} - w_{t}

I am interested in finding out

\Delta_w / w_{t}

Is there a way to do that?