How to parallelize if my network can not batch?

There is pytorch.multiprocessing but I haven’t taken the time to understand how it works yet. If you find yourself constrained to giving batches to your model, you could use batches of size 1, and alter your model to squeeze the batch dimension when necessary.

Personally I would try using joblib as in the following example. Again, I have not yet tried it, but I think it would work.