Distributed training on multiple node

Hi,

Currently I am working on two multi node configuration with 4 GPU each on the imagenet example provided by pytorch with the ip of:

root@111.111.0.1
root@111.111.0.2

Its there a way i can distributed data parallelism using Pytorch?