I am encountering the same error when fine tuning pre trained model when splitting processing between 2 gpus using DataParallel.
Have you found a solution?
I am encountering the same error when fine tuning pre trained model when splitting processing between 2 gpus using DataParallel.
Have you found a solution?