Hi,
I would like to train a Llama 2 7b based on a singular RTX 4070 GPU with a small dataset by running auto train command locally:
autotrain llm --train --project-name my-llm --model meta-llama/Llama-2-7b-hf --data-path timdettmers/openassistant-guanaco --use-peft --quantization int4 --lr 2e-4 --batch 2 --epochs 3 --trainer sft
But the esitimation is still like this after running almost half a day.
1%|▌ | 47/6393 [11:58:31<1612:04:01, 914.50s/it]
Is there any approach or parameters changes which can make it quicker?
Any suggestions or helps will be highly appreciated.
K