So I built PyTorch and torch_xla following these discussions and I can import torch and torch_xla but I can’t get any devices to work. When I call xla_device() I get an error saying “Missing XLA configuration”. I want to run on GPU ideally. If I call xla_device(devkind=‘GPU’) the call just hangs (no clue if thats a valid device kind but I thought I’d give it a try).
nvidia-smi returns my desired results, nvcc --version returns 10.1, however torch.cuda.is_available() returns false.
I’m not really sure how to debug further here. It seems like everything built correctly but something in my environment isn’t configured correctly. I’m not sure how to find out what that is however. Any thoughts?