[Resolved] Cuda runtime error

I trained my model on the 2 different GPUs(using Tensor.cuda()) and saved the parameters. Then I want to load these parameters to the CPU model, but I got this error: cuda runtime error (10) : invalid device ordinal at torch/csrc/cuda/Module.cpp:84
How can I solve this?


Found answer in this topic