I am trying to experiment models for RAG using my official documents. For this I would like to run model on my local machine.
But as I am new to LLM world, I keep hitting roadblock because some models have specific requirements and I don’t find it explicitly mentioned on model page. For e.g. I was trying to run LlaMa 2 on my m1 mac, but then to realize that I would need CUDA suitable GPU for it to run.
Is there any way I can understand what is the minimum system requirements from hardware and software perspective to run a given LLM model ?