Introduction for the Accelerate
library says I have to be willing to write a forward loop (forgoing Trainer). Is there a way for me to enable DDP training while continuing using Trainer?
Replacing _get_train_sampler
with _get_eval_sampler
looks like a much more elegant solution, thank you!