Options for mobile inference

How can optimize a PyTorch model so I can run it on Android, without needing to convert the model to TorchScript?

For now there aren’t all that many options, you can convert your model to torchscript, ONNX or tflite. But regardless of what you pick there will be some conversion process because phones don’t typically support python natively (iOS certainly doesn’t but perhaps with some workarounds it’s possible to do this in android)