Hi all, I’m using this .py file (sentence-transformers/examples/unsupervised_learning/MLM at master · UKPLab/sentence-transformers · GitHub) to carry out some unsupervised training.
It’s returning a message saying: ‘We strongly recommend passing in an attention_mask
since your input_ids may be padded. See Troubleshoot’
I’m a little unfamiliar with this process, but my guess is that it may be due to the fact that my training data text file has every row varying in length.
How would I adapt this to add in an attention mask? Thanks!