I am using v1.0.1/CUDA 10.1 on a 2080. glibc is 2.28.
I read that libtorch is multithread safe but I cannot make it work correctly in a MT C++ app. I have 2 trainers each one in its own thread, no global var, no static var, no shared var… but it always ends in a deadlock.
When deadlocked both threads are near cuLaunchKernel. Is there something I need to do ? Do I have to initialize a context or something?
I also tested with master and a nightly of libtorch but it ends the same.