How was LlamaForSequenceClassification Pretrained

Hello!

We are relatively new to HuggingFace, and we are trying to access some form of document (not token) embedding for LLaMA or other LLMs. We noticed that LlamaForSequenceClassification has two outputs and uses the first token to represent hidden states (so perhaps the document embedding we are looking for?).

We could not find on which task LlamaForSequenceClassification was fine-tuned anywhere. We’d appreciate if you can point us in the right direction.

1 Like