As the title suggest, every time I fine-tune my model, I get no actual result, I just get the last word repeated multiple times. The model I’m fine-tuning is RoFormer, and here is an example of what is happening:
>> prompt = "Somatic hypermutation allows the immune system to"
>> generator(prompt)
[{'generated_text': 'Somatic hypermutation allows the immune system to to to to'}]
Here is a bit of details:-
- I tried multiple datasets.
- I tried following the official documentation for training a Casual LM.
- The training, and validation losses are also really small.
- The base language is Chinese.
What am I missing?