How to run an end to end example of distributed data parallel with hugging face's trainer api (ideally on a single node multiple gpus)?

ok this is the command:

torchrun --nproc_per_node 2 my_script.py
1 Like