This might be better suited to #beginners, but it is definitely #transformers-specific… I’m new to and wanted to try model adaptation of one of the Helsinki-NLP MT models using fine-tuning.
I’ve created a DatasetDict with train, dev and test, and managed to load the pre-trained model and run a trainer for one epoch. My corpus is very small (<1k segments), so my expectation is that it would have little impact on the baseline model.
However, when I use the locally-saved config to translate, the results look more like the output of a model just starting its training. Is this expected, or am I doing something terribly wrong?