Runing Multiple trainings (or inference) in parallel in a single GPU

Is it possible to run Multiple trainings of the same deep learning model (pytorch) in parallel in a single GPU ? and How please ?

Yes, this would be theoretically possible if:

  • enough memory and compute resources are free on the GPU
  • if you are using different streams or generally different processes
  • if the CPU is not blocking the execution.