Which data parallel does trainer use? DP or DDP?

It depends if you launch your training script with python (in which case it will use DP) or python -m torch.distributed.launch (in which case it will use DDP).

4 Likes