Extracting token embeddings from pretrained language models

Thank you for the very clean answer, Omar!

May I ask you to elaborate on the differences between your answer and the following ones?

If I understood well, in practice, all the questions are related to the same request: the word embedding extraction from pre-trained models. If so, what is the best practice among all the reported solutions?

Sorry if I missing something and thank you for your clarification. :slightly_smiling_face: