PyTorch Forums
Should we split batch_size according to ngpu_per_node when DistributedDataparallel
distributed
junb
November 18, 2021, 10:01am
19
Good call, I will have to remember that! Thanks for your quick answer.
1 Like
How to scale/warmup the learning rate for large batch size?
show post in topic