How to overlap coping data to/from the GPU with performing computations on the GPU?

Hi, I am new to pytorch.

I want to know how to overlap copying data to/from the GPU, with performing computations on the GPU ?

Thanks very much!

Hi,

You can use the non_blocking argument to either the copy_ or to() or cuda() functions to be able to do that.