Fasterrcnn_resnet50_fpn parallelization warning

When training PyTorch built-in model fasterrcnn_resnet50_fpn on a custom dataset using nn.DataParallel, I received the following warning:

UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; will instead unsqueeze and return a vector.
  warnings.warn('Was asked to gather along dimension 0, but all '

By this post, I realized that the problem is, fasterrcnn_resnet50_fpn is using loss function inside the forward function.

So what should I do to fix this? I tried to ignore it and divide the final loss by num_workers. However, it seems that the model is not parallelized and all work was done by GPU 0 (maybe this is another irrelevant problem).

You probably already found it, but this model won’t work with nn.DataParallel. Train with Distributed like here:

1 Like