I encounter errors when I attempt to use a batch size that is not divisible by 3. Is there an option that allows the last GPU to handle the remaining part of the batch?
Example
batch_size=32
can be divided as
GPU1 : 10 (batch_size//3)
GPU2 : 10 (batch_size//3)
GPU3 : 12 (batch_size//3+r)