OSError: meta-llama/Llama-2-7b-chat-hf is not a local folder and is not a valid model identifier listed on ‘https://huggingface.co/models’
If this is a private repository, make sure to pass a token having permission to this repo with use_auth_token
or log in with huggingface-cli login
and pass use_auth_token=True
.
Hi there! Did you follow the error instructions? (either logging in or passing the token)
Yes i have followed the error and i got access to the model through meta,hugginface also