I am using huggingface’s Trainer with --bfl16 flag enabled and deepspeed enabled. However, I want to force float32 for a specific layer. How to do it?
--bfl16
Hmm… Mixed precision training?
Or similar to this issue…?