Is it possible to bypass the tokenizer and directly provide the input embeddings to train the BERT model?
You can just create a subclass of the model that you want and modify its forward pass.
You can also feed
input_embeds instead of
input_ids to your mdel.