How to use an `optimizer` to update non-differentiable parameters?

y_q_diffable is y_q ( = y + y_q - y) for the forward. But during backwards, the gradients propagate as if y_q_diffable were y.

Best regards

Thomas

2 Likes