I was thinking about that for Polish. Few thoughts:
-
I assume you mean fine-tuning with LoRA and not the llama2 itself
-
If Llama2 has not been trained on significant portion of a foreign language then fine-tuning a chat model with LoRA might not help a lot - I recently read a paper that says exactly this, don’t remember where.
-
People still doing this, not sure how successful, for example: davidkim205/komt-Llama-2-13b-hf-lora · Hugging Face
-
If you want to train base model (a not a chat model) to include more Bulgarian, then you have to go through RLHF, which is difficult and costly.