I’ve worked with the guide from peft docs only slightly adjusting the preprocessing function to resolve one bug and substituting phi-2 for the original model. The trained model produces repeating output, for example:
[‘Tweet text : @NYTsupport i have complained a dozen times & yet my papers are still thrown FAR from my door. Why is this so hard to resolve? Label : complaintaintcomplaintaint’]
I’m wondering what could be the cause and if there is anything that can be done about it?
I found out where the problem was. Turns out the eos token set in config doesn’t actually influence what’s going on during inference. It has to be set as an argument of the generate method. [0] All credit for this solution should go to Eugenio-Schiavoni who first wrote about it on github.com