Llama3 Fine-Tuning Consultation

I am a college student racking my brain over my thesis project. My task is to fine-tune Llama3.1-8b using 50,000 QA pairs.

The QA data comes from the following source:
https://huggingface.co/datasets/aigrant/medical_health
I am extracting only the instruction, input, and output fields and organizing them into a .json file in the following format:
[
{
“instruction”: “了解過敏性鼻炎”,
“input”: “過敏性鼻炎的趨勢和症狀”,
“output”: “過敏性鼻炎患者逐年增加,孩童約有一半會有鼻子問題,成人約20%有症狀。症狀包括鼻癢、打噴嚏、流鼻水、鼻塞等。”
},…
We plan to use this on Ollama. I am fine-tuning with Unsloth and attempting to run it on either Colab or my local machine. However, over the past three months, no matter how I fine-tune it—despite following various online tutorials—the .gguf file I generate not only fails to work on Ollama (create fails) but also doesn’t work on other software. The responses turn into gibberish or something resembling multiple-choice questions.

If anyone can guide me or suggest an alternative solution, I would greatly appreciate it.

Currently, I am using an RTX 3050 Ti Laptop GPU, and our lab has an RTX 3090 available for use.

1 Like

Was it working properly before it was converted to GGUF?