I fine-tuned a falcon-7b model and called trainer.train()
using SFTTrainer
from huggingfaceâs trl
package.
Calling trainer.model.push_to_hub('hub_name')
pushes three files to the hugginface repositoryâa README.md
, an adapter_config.json
, and adapter_model.bin
. Iâm having trouble finding any documentation that describes how to use these file formats.
At first I found this huggingface page, âUsing Adapter Transformers at Hugging Face.)â, and it says to use the class âAutoModelWithHeadsâ. However, this class wonât load after downloading the adapter-transforms
pip package on Google Colab, I think itâs deprecated? Also, calling
adapter_name = model.load_adapter(ADAPTER)
model.active_adapters = adapter_name
with AutoModel results in this error:
AttributeError: âRWForCausalLMâ object has no attribute âload_adapterâ
How can you apply adapter_model.bin files? My code was based on the official Falcon-7b huggingface tutorial, where they link a Google Colab notebook on that page.