How to re-use same gpu memory for each forward pass?

Anyway, I figured out how to do this in a way similar to torch. Just create a buffer tensor t in init(), then in forward pass, wrap it with Variable(t)