FineTuning a LLM-Model

I want to fine tune a small LLM Model like Llama etc… to chat with arxiv publications.

Example questions:

  • Show me all publication issued in 2022
  • Show me all author names issues publications in 2022

Question

  • Is it even possible and how to start?
  • Can I restrict a LLM after Tuning only to the new data (arxiv publications)
1 Like

Thank you on having me.

  1. Your idea is possible.

It is possible task. In my research team fine tune the model using new dataset esp.)medical dataset. So it is not a problem to fine tune the model on the arxiv publications if the dataset is correct.

  1. Arxiv dataset

In my opinion, you should use langchain framework instead of fine tuning your model. Fine tuning might be right way to build your own arxiv assist. But it costs too much. So I recommend you using langchain. Langchain support the arxiv module.

1 Like