A document question regarding batch inference example with TorchServe

Hey I am looking into how to use TorchServe to do batch inference.

The example in this document is supposed to demonstrate how to do batch inference, but it seems to have only passed one image.

Is it an unintentional example or expected? Thank you!

I believe the example shows that single images are still passed to TorchServe, but a batched inference will be used if multiple requests would be received in the specified max_batch_delay time window. At least this is how I understand the example, but others can of course correct me.