I haven’t tried it but you could convert your model to ONNX before deploying to FPGA Deploy ML models to FPGAs - Azure Machine Learning | Microsoft Docs
But in general there’s easier ways of increasing inference time than using FPGA Inference extremely slow on AWS (CPU)? - #2 by marksaroufim