Cannot Download Dolly Due to 'OSError: Distant resource does not seem to be on huggingface.co (missing commit header).'

I am getting this OS error every time I try to download the databricks/dolly-v2-3b model, even though I could download it last week, and nothing changed in my environment at all. I keep getting the same recommendations about upgrading or downgrading transformers, requests, huggingface_hub. Here’s my code.

#imports
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModelWithLMHead
import re
import pandas as pd
pd.set_option('display.max_colwidth', None)
pd.set_option('display.max_rows', None) 
from sklearn.model_selection import train_test_split
from sklearn.metrics.pairwise import cosine_similarity
from sklearn.feature_extraction.text import TfidfVectorizer
import os
os.environ['CURL_CA_BUNDLE'] = ''

#load model
tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-3b", padding_side='left')
model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-3b")

and here is the trace to the error

---------------------------------------------------------------------------
OSError                                   Traceback (most recent call last)
~\AppData\Local\Temp\ipykernel_4476\1259564569.py in <module>
      1 #load model
----> 2 tokenizer = AutoTokenizer.from_pretrained("databricks/dolly-v2-3b", padding_side='left')
      3 model = AutoModelForCausalLM.from_pretrained("databricks/dolly-v2-3b")

c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\models\auto\tokenization_auto.py in from_pretrained(cls, pretrained_model_name_or_path, *inputs, **kwargs)
    641 
    642         # Next, let's try to use the tokenizer_config file to get the tokenizer class.
--> 643         tokenizer_config = get_tokenizer_config(pretrained_model_name_or_path, **kwargs)
    644         if "_commit_hash" in tokenizer_config:
    645             kwargs["_commit_hash"] = tokenizer_config["_commit_hash"]

c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\models\auto\tokenization_auto.py in get_tokenizer_config(pretrained_model_name_or_path, cache_dir, force_download, resume_download, proxies, use_auth_token, revision, local_files_only, subfolder, **kwargs)
    498         _raise_exceptions_for_missing_entries=False,
    499         _raise_exceptions_for_connection_errors=False,
--> 500         _commit_hash=commit_hash,
    501     )
    502     if resolved_config_file is None:

c:\Users\samuel.cannon\AppData\Local\anaconda3\envs\decorations_generator_venv\lib\site-packages\transformers\utils\hub.py in cached_file(path_or_repo_id, filename, cache_dir, force_download, resume_download, proxies, use_auth_token, revision, local_files_only, subfolder, repo_type, user_agent, _raise_exceptions_for_missing_entries, _raise_exceptions_for_connection_errors, _commit_hash)
    427             resume_download=resume_download,
    428             use_auth_token=use_auth_token,
--> 429             local_files_only=local_files_only,
...
-> 1214                 raise OSError("Distant resource does not seem to be on huggingface.co (missing commit header).")
   1215 
   1216             # Etag must exist

OSError: Distant resource does not seem to be on huggingface.co (missing commit header).

well this has been really helpful lol