LLAMA for MCQA dataset

Hi, I am fine-tuning LLAMA on a multiple-choice question-answering (MCQA) dataset. During the training phase, would it be a good approach to trim the model’s output head to just four tokens corresponding to the answer options, so that during the generation phase, the model is constrained to generate only the labels? Are there any alternative strategies I could consider for achieving this?

1 Like