Does the minibatch size divides into number of GPUs when using nn.dataparallel?
I had the minibatch size of 16 and when I use nn.dataparallel with 4 GPUs,
the actual initial input shape which goes through the model (printed it in the beginning of foward function) was like this
[4, channel, width, height]
[4, channel, width, height]
[4, channel, width, height]
[4, channel, width, height]
Is this because of the minibatch size is divided into each 4 GPUs and compute parallel?