Hi, i am attempting to train a CodeLlama-34b-v2 model on a custom dataset of front end code. I have tried to do this using GCPs Vertex AI; however, the integration with huggingface and GCP resources is not that intuitive. Does anyone have any experience with training this model on larger datasets of code?
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
How to fine-tune a pretrained LLM on custom code libraries? | 3 | 6748 | April 26, 2025 | |
Fine-tuning CodeLlama for Multi-File Code Generation in a Private Repository | 10 | 7872 | October 23, 2024 | |
How to Efficiently Fine-Tune Models on Custom Datasets with Limited Resources? | 0 | 114 | July 10, 2024 | |
Finetune GPT-J on custom dataset | 0 | 2804 | January 18, 2022 | |
Help with Training a Custom Model using Hugging Face Transformers | 0 | 28 | October 11, 2024 |