Thank you!
The linked notebooks seem to do a lot of boilerplate work that is now handled in the seq2seq finetune.py script?
After spending a few hours reading through code, it seems like I should be able to finetune by just
- updating the tokens as you’ve described
- creating a dataset (train.source, train.target, etc)
- running finetune.sh with those as parameters
Is this correct?