Is it possible to run Multiple trainings of the same deep learning model (pytorch) in parallel in a single GPU ? and How please ?
Yes, this would be theoretically possible if:
- enough memory and compute resources are free on the GPU
- if you are using different streams or generally different processes
- if the CPU is not blocking the execution.