Then, I would like to fine-tune this model and safe it. Afterwards, I would like to load the fine-tuned model with out load_in_8bit so that I can run it on Windows (bitsandbytes library which is needed for load_in_8bit is not supported on Windows).
Is this possible or do I need to run a model fine-tuned with load_in_8bit always with this flag?
Do you know if it is possible to train using 8-bit optimizer from bitsandbytes but save it “normally” (without 8-bit) so that I can load the model without bitsandbytes?
Yes, it’s only the optimizer state which is in int8, not the model state. You just need to replace your regular optimizer by bnb.optim.Adam8bit(....)