DataParallel for Mixed GPUs

Could pytorch effectively use setup in server with different GPUs, for example 2x 2080 Ti + 2x 3080 Ti for the one distributed learning process? Or some drawbacks may occur?

We had similar posts in this forum, which you could search for, and the main drawback would be that the slower GPUs would potentially create the bottleneck in your setup.