Models on Different GPUs

I think @apaszke’s answer on a previous post here helps too, and is also closer to what I’m looking for although I’m not looking to split over a model [Model parallelism in pytorch for large(r than 1 GPU) models?](http://Model Parallelism in Pytorch)