Hi, I’m trying to fine-tune GPT NeoX 20B using LoRA and peft - the process goes great, takes about 12 hours on my dataset, training loss is acceptable… But when it is finished, the adapter_model.bin file is very small for some reason (443 bytes) when it should have at least a few MB. When I load the adapter, the model gives unexpected outputs, which shouldn’t happen. Here is the script: alpaca-lora_gpt_neox_20b/finetune.py at main · satani99/alpaca-lora_gpt_neox_20b · GitHub (base_model is not specified, but I specified it before I began the fine-tuning process).
I might be wrong, but I think the problem is in the lora_target_modules, where the module is incorrectly specified. I tried printing them after loading the model, and there are like 120 target modules, so which one should I target, all of them?
If anyone knows what the problem is, please let me know.