Hello, I’m trying to compare how inference time getting faster by reducing FLOPs through changing input sizes.
However, torchvision resnet 50 does not seems to get faster by size reduction.
So I test on github (address below) resnet 50 that gets faster by smaller input.
What makes the difference?
(Besides time takes almost 4 times bigger in github code on same 300x300 size)
x-axis showing the length of the input, test by Tensor shape (1, 3, x , x)
y-axis showing inference time
The line with orange means a theoretical possible reduction by FLOPs change.
The blue line is experimental num.
github code (with no Fully connected layer and avgpool layer)
torchvsion code (with changing Fully connected layer and avgpool layer to Idenetity layer)