I am getting this OS error every time I try to download the databricks/dolly-v2-3b model, even though I could download it last week, and nothing changed in my environment at all. I keep getting the same recommendations about upgrading or downgrading transformers, requests, huggingface_hub. Here’s my code.
#imports
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModelWithLMHead
import re
import pandas as pd
pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_rows', None)
from sklearn.model_selection import train_test_split
from sklearn.metrics.pairwise import cosine_similarity
from sklearn.feature_extraction.text import TfidfVectorizer
import os
os.environ['CURL_CA_BUNDLE'] = ''
#load model
tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-3b", padding_side='left')
model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-3b")
and here is the trace to the error
---------------------------------------------------------------------------
OSError Traceback (most recent call last)
~\AppData\Local\Temp\ipykernel_4476\1259564569.py in <module>
1 #load model
----> 2 tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-3b", padding_side='left')
3 model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-3b")
c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\models\auto\tokenization_auto.py in from_pretrained(cls, pretrained_model_name_or_path, *inputs, **kwargs)
641
642 # Next, let's try to use the tokenizer_config file to get the tokenizer class.
--> 643 tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs)
644 if "_commit_hash" in tokenizer_config:
645 kwargs["_commit_hash"] = tokenizer_config["_commit_hash"]
c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\models\auto\tokenization_auto.py in get_tokenizer_config(pretrained_model_name_or_path, cache_dir, force_download, resume_download, proxies, use_auth_token, revision, local_files_only, subfolder, **kwargs)
498 _raise_exceptions_for_missing_entries=False,
499 _raise_exceptions_for_connection_errors=False,
--> 500 _commit_hash=commit_hash,
501 )
502 if resolved_config_file is None:
c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\utils\hub.py in cached_file(path_or_repo_id, filename, cache_dir, force_download, resume_download, proxies, use_auth_token, revision, local_files_only, subfolder, repo_type, user_agent, _raise_exceptions_for_missing_entries, _raise_exceptions_for_connection_errors, _commit_hash)
427 resume_download=resume_download,
428 use_auth_token=use_auth_token,
--> 429 local_files_only=local_files_only,
...
-> 1214 raise OSError("Distant resource does not seem to be on huggingface.co (missing commit header).")
1215
1216 # Etag must exist
OSError: Distant resource does not seem to be on huggingface.co (missing commit header).