is there an equivalent of Tensorflow Serving in Pytorch? More specifically, automated inference server that handles batching requests to maximize performance, switching models, running experimental models and recording performance…
tensorflow serving: https://www.tensorflow.org/serving/