I took. for rent a server with 8 GPU and 8 CPU. I want to use DistributedDataParallel. Please give me a simple example of how to do this. I found a lot of different examples, but I do not fully understand them.
Hey @slavavs
This is the minimum example.
For more advanced usages, please see this overview.
BTW, could you please add a “distributed” tag for questions related to distributed training, so that people work on that can get back to you promptly. Thx!