Following the blog, I was able to run inference on my gpu server for meta-llama/Llama-2-7b-chat-hf as has been done in the tutorial. (Have been granted access to Llama-2 already)
Now I wish to run inference on meta-llama/Llama-2-13b. I keep getting the error:
“OSError: meta-llama/Llama-2-13b does not appear to have a file named config.json. Checkout ‘https://huggingface.co/meta-llama/Llama-2-13b/main’ for available files.”
Checking the mentioned repo for config.json, I see there actually isn’t one for meta-llama/Llama-2-13b (and most other Llama2 models except meta-llama/Llama-2-7b-chat-hf).
Could the missing file(s) it be added please? Alternatively, if there’s another way to do a run the model privately, it would be great. Thanks!