Is there a way to stop the process of the inference on the GPU?
Description: I load a object detection model onto the GPU. After that I want to run the inferences on the GPU. Is there a possibility to stop the inference process on the GPU (for example after a specified time)? But the model should remain on the GPU and a reload of the model onto the GPU should be prevented.