Distributed training on just cpu on a single node

How to use ddp_backend=“gloo” for trainer for running distributed training on multiple docker containers on a single node?

  1. Do I still need to initialize with dist.init_process_group within the script?
  2. What would be the world size while running multiple containers on a single node?