while running multiple PyTorch scripts Using DataParallel, only the first one is distributed among two GPUs, rest are loading 1st GPU memory.
Cant all the program distribute themselves like 1st one?
If they can what’s the process?
I am just loading the model and using
model = ConvTasNet(args.N, args.L, args.B, args.Sk, args.H, args.P, args.X, args.R,
args.C, norm_type=args.norm_type, causal=args.causal,
mask_nonlinear=args.mask_nonlinear)
print(model)
if args.use_cuda:
model = torch.nn.DataParallel(model)
model.cuda()
Same procedure for all scripts.