I wanted to use Masked language modelling as method to pretrain text on domain specific tasks. But for open source LLMs like MPT - They only work on autocasualLM. Is there a way to do MLM on MPT 30B?
I wanted to use Masked language modelling as method to pretrain text on domain specific tasks. But for open source LLMs like MPT - They only work on autocasualLM. Is there a way to do MLM on MPT 30B?