Hi all!
I’ve trained my 80million parameter -model successfully on the same system.
Few weeks later, I decreased my model network to 18 million parameter but now it is saying that
CUDA out of memory
How is this even possible?
Thanks!
Hi all!
I’ve trained my 80million parameter -model successfully on the same system.
Few weeks later, I decreased my model network to 18 million parameter but now it is saying that
CUDA out of memory
How is this even possible?
Thanks!
Hi Furkan,
Two potential culprits jump to mind here:
Without more detail, it’s hard to give a clear answer, but perhaps you could investigate those two hypotheses and see if anything pops up.
Best,
Andrei
Thanks Andrei. Your suggestions are probably the reason of this issue. I actually updated the torch to 1.13.1. Then, the model was run without a hitch.
Best,
Furkan