Hi, I am fine-tuning LLAMA on a multiple-choice question-answering (MCQA) dataset. During the training phase, would it be a good approach to trim the model’s output head to just four tokens corresponding to the answer options, so that during the generation phase, the model is constrained to generate only the labels? Are there any alternative strategies I could consider for achieving this?
Related topics
| Topic | Replies | Views | Activity | |
|---|---|---|---|---|
| Finetuning llama for classification | 2 | 1113 | January 21, 2025 | |
| Finetuned model generating test label exactly | 0 | 466 | October 15, 2020 | |
| Question-Answering/Text-generation/Summarizing: Fine-tune on multiple answers | 8 | 5326 | November 20, 2021 | |
| Labels in language modeling: which tokens to set to -100? | 1 | 3501 | November 30, 2020 | |
| Impossible questions when finetuning QA models | 0 | 294 | November 19, 2021 |