With ROCm 7.0 just released (and probably soon supported by PyTorch), and new interesting AMD Radeon AI PRO R9700 AI TOP 32 GB VRAM on the market – which seems like a very tempting proposition – I’m curious how far ROCm has and PyTorch has come for a smooth inference and training.
So far, I only ever worked with Nvidia cards and CUDA, and never had any (real) issues. However, Nvidia cards with 32 GB VRAM are prohibitively expensive. I’ve tried to read up on this, but could not find a good consensus. Many just say to stick with Nvidia/CUDA, others say ROCm and Python works just fine, and others say it works fine for inference but training can be tricky.Any strong opinions in this forum?
Any comments before committing to a large purchase are welcome :).
Yes, I probably should have added that I’m on Linux (Ubuntu) which seems to make life easier than Windows with ROCm (again, simply based on what I’ve read online).
Yea that’s right. For whoever stumbles on this, I have tested all my Pytorch-CUDA code extensively on a Fedora system with dual 7900XTX with that pyproject.toml file.