Unusual performance results for ResNet50

(Sebastian) #1

I have found some very unusual behavior when comparing time of inference run for ResNet50 in image classifier for various backends. I’d very much appreciate any comment why such thing might have happened.

My line looks like this:

./bin/image-classifier <testImages> -use-imagenet-normalization -image_mode=0to1 -m=resnet50 -model_input_name=gpu_0/data -<cpu|interpreter|opencl> -time

Where <testImages> is directory with 1/16/32/64 images depending on the tested batch size. To have a bit clearer picture I have used measured time to calculate images/second value.

And now the strange part…

For OCL I can see logarithmic increase, but for all other backends value is constant. It completely doesn’t matter how big batch size I throw at it, images/second value is the same (+/- small noise).

On generated dot file everything looks reasonable, that is batch size is propagated.

Have you also experienced similar behavior? Or maybe I’m just doing my measurements wrong?

(Sebastian) #2

Below you can see chart presenting number of images/second with increase of batch size for each of backends.

chart

(Jordan Fix) #3

Hi Sebastian, I’m assuming this is because the OpenCL backend takes advantage of parallelism across the multiple images in the batch, whereas our CPU and Interpreter do not and are always running on a single CPU core. So I would expect to always see about the same images/second for the CPU and Interpreter, whereas the OpenCL backend will benefit from this parallelism up until whatever device you’re using has all of its parallel resources exhausted, at which point images/second would plateau. This appears to be what you are seeing here.

1 Like
(Sebastian) #4

Thanks Jordan! That makes a lot of sense now. I am almost sure that I read about that limitation. Apparently I must have forgot.

1 Like
(Jordan Fix) #5

By the way, there have been some discussions about multi-threaded CPU kernels – you can follow along in GH issue #1749.