Hi, I have been trying to train and test faster_rcnn where I want to train and test pytorch image with the following parameter
model =torchvision.models.detection.fasterrcnn_resnet50_fpn(pretrained=True, min_size = 200,max_size = 4000)
but it seems image is resized upon inference eventhough it shouldn’t be inference image sizes are : (912,912), (1824,1824), (3648,3648)
I know that image is resized upon inference since the time and GPU RAM usage doesn’t make sense (all size are similar even if I set the min_size = 200, and max_size = 4000, but if I change min_size and max_size close to the original size like (3648, 3648) the time and RAM usage will be different which is why I draw conclusion that the image is resized)
problem is, because I felt like the generalized transform is so whimsical I don’t even know if my train is the right size/ model learn the right information.
Although now that I think about it I didn’t use dataloader while inferencing (straight using util_dataset) does it affect the inference image size in any way?