Hello,
I am trying to finetune Llama 2 7B to enhance its function calling abilities. I am not sure what I am doing wrong, but the model is not really converging while training and the performance, hence, isn’t good at all.
I am using this Jupyter Notebook and tried it with and without quantization.
These are the training and eval/test datasets. I made them, and they are designed to learn the ‘get_current_weather’ function.
As my datasets are small, I picked higher values for r and lora_alpha
r=256, lora_alpha=512
Does anyone have any ideas on how to improve the model? Am I doing something wrong?
I am very new to the topic of finetuning LLMs. So, any feedback or help would be very much appreciated.