I am trying to launch a container that calls the HuggingFace hub but there seems to be a connection or cache error. why is this happening?
$ sudo docker compose up
...
h2ogpt | Traceback (most recent call last):
h2ogpt | File "/workspace/generate.py", line 16, in <module>
h2ogpt | entrypoint_main()
h2ogpt | File "/workspace/generate.py", line 12, in entrypoint_main
h2ogpt | H2O_Fire(main)
h2ogpt | File "/workspace/src/utils.py", line 65, in H2O_Fire
h2ogpt | fire.Fire(component=component, command=args)
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/fire/core.py", line 141, in Fire
h2ogpt | component_trace = _Fire(component, args, parsed_flag_args, context, name)
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/fire/core.py", line 475, in _Fire
h2ogpt | component, remaining_args = _CallAndUpdateTrace(
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace
h2ogpt | component = fn(*varargs, **kwargs)
h2ogpt | File "/workspace/src/gen.py", line 1664, in main
h2ogpt | model=get_embedding(use_openai_embedding, hf_embedding_model=hf_embedding_model,
h2ogpt | File "/workspace/src/gpt_langchain.py", line 461, in get_embedding
h2ogpt | embedding = HuggingFaceEmbeddings(model_name=hf_embedding_model, model_kwargs=model_kwargs)
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/langchain_community/embeddings/huggingface.py", line 65, in __init__
h2ogpt | self.client = sentence_transformers.SentenceTransformer(
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/sentence_transformers/SentenceTransformer.py", line 87, in __init__
h2ogpt | snapshot_download(model_name_or_path,
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/sentence_transformers/util.py", line 491, in snapshot_download
h2ogpt | path = cached_download(**cached_download_args)
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 118, in _inner_fn
h2ogpt | return fn(*args, **kwargs)
h2ogpt | File "/h2ogpt_conda/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 770, in cached_download
h2ogpt | raise LocalEntryNotFoundError(
h2ogpt | huggingface_hub.utils._errors.LocalEntryNotFoundError: Connection error, and we cannot find the requested files in the cached path. Please try again or make sure your Internet connection is on.
^X^CGracefully stopping... (press Ctrl+C again to force)
Aborting on container exit...
[+] Stopping 1/1
✔ Container h2ogpt Stopped 0.5s
canceled
Here is my docker-compose.yml
:
version: '3'
services:
h2ogpt:
image: gcr.io/vorvan/h2oai/h2ogpt-runtime:latest
container_name: h2ogpt
shm_size: '2gb'
environment:
- ANONYMIZED_TELEMETRY=False
- HF_DATASETS_OFFLINE=1
- TRANSFORMERS_OFFLINE=1
volumes:
# - $HOME/.cache:/workspace/.cache
- ./data/models:/workspace/models:ro
- ./data/save:/workspace/save
- ./data/user_path:/workspace/user_path
- ./data/db_dir_UserData:/workspace/db_dir_UserData
- ./data/users:/workspace/users
- ./data/db_nonusers:/workspace/db_nonusers
- ./data/llamacpp_path:/workspace/llamacpp_path
- ./data/h2ogpt_auth:/workspace/h2ogpt_auth
ports:
- 7860:7860
restart: always
deploy:
resources:
reservations:
devices:
- driver: nvidia
count: all
capabilities: [gpu]
command: >
/workspace/generate.py
--base_model=mistralai/Mistral-7B-Instruct-v0.2
--hf_embedding_model=intfloat/multilingual-e5-large
--load_4bit=True
--use_flash_attention_2=True
--score_model=None
--top_k_docs=10
--max_input_tokens=2048
--visible_h2ogpt_logo=False
--dark=True
--visible_tos_tab=True
--langchain_modes="['UserData', 'LLM']"
--langchain_mode_paths="{'UserData':'/workspace/user_path/sample_docs'}"
--langchain_mode_types="{'UserData':'shared'}"
--enable_pdf_doctr=off
--enable_captions=False
--enable_llava=False
--use_unstructured=False
--enable_doctr=False
--enable_transcriptions=False
--enable_heap_analytics=False
--use_auth_token=hf_XXXX
--prompt_type=mistral
--pre_prompt_query="Use the following pieces of informations to answer, don't try to make up an answer, just say I don't know if you don't know. Answer in the following language: french"
--prompt_query="Cite relevant passages from context to justify your answer."
--use_safetensors=False --verbose=True
networks:
- h2ogpt-net
networks:
h2ogpt-net: