meta-llama/Meta-Llama-3.1-8B-Instruct
I follow the tutorial and it show the error.
‘’’
Cannot access gated repo for url https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct/resolve/main/config.json.
Access to model meta-llama/Meta-Llama-3.1-8B-Instruct is restricted. You must be authenticated to access it.
‘’’
Q2.
I download the Meta-Llama-3.1-8B-Instruct by git clone
How to load it with the code?
I dont see any solution?
pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)
             
            
              
              
              
            
            
           
          
            
            
              hi @alice86
You need to agree the license agreement to access this model. You can find it in the model card.
FAQ
You will need this one:
             
            
              
              
              
            
            
           
          
            
            
              Q1
I already accept the license.
And the official don’t show any method to download it.
It just direct use
pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)
Q2
It is different question.
If I download the model and how to load  model  without token? (already download model)
Because I already download the model.
I expect the result is
python test.py    => it will automatically download model with token
python test.py  -m ./model/llama3 => It will use the model that I already download without any token.
In fact, I dont see any code for   python test.py  -m ./model/llama3
             
            
              
              
              
            
            
           
          
            
            
              hi @alice86
About your first question:
Did you create an access token? Did you add a permission for relevant repository?
You can check from Hugging Face – The AI community building the future..
(Edit permissions → Repositories permissions)
You need to run something like this:
pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
    token="hf_xxxxxxx",
)
             
            
              
              
              
            
            
           
          
            
            
              About your second question:
You need to call the pipeline with model='./path_for_local_model':
pipeline = transformers.pipeline(
    "text-generation",
    model='./path_for_local_model',
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto",
)