I am training a transformer based model on a GPU and get an out of Memory error. This is the error message:
CUDA out of memory. Tried to allocate 578.00 MiB (GPU 2; 10.76 GiB total capacity; 2.79 GiB already allocated; 6.55 GiB free; 3.24 GiB reserved in total by PyTorch)
The model runs fine on a different GPU.
I would truly appreciate any help!
Thank you all