“Use this model”->Ollama: can't pull model with Q4

Can’t pull Q4 model from IlyaGusev/saiga_llama3_8b_gguf · Hugging Face via “Use this model”->Ollama button.
All Q2, Q4, Q8, f16 exist in repository, I can download files, but “Use this model”->Ollama only works with Q2_K, Q8_0, F16. It worked month ago.

By default the model field is empty, but if I copy I will get: “ollama run hf.co/IlyaGusev/saiga_llama3_8b_gguf:Q4_K_M


If I run ollama pull hfco/IlyaGusev/saiga_llama3_8b_gguf:Q4_K_M output is “The specified tag is not available in the repository. Please use another tag or latest”

If I change tag and run ollama pull hfco/IlyaGusev/saiga_llama3_8b_gguf:Q4_K output is “The specified tag is not a valid quantization scheme. Please use another tag or latest”

Is this HF issue?

1 Like

I’m not sure if this is an HF issue or an Ollama issue…

Either way, it’s an issue.

2 Likes

I’m not sure if this is a bug, but I noticed an inconsistency in the file type labels for these GGUF models:

  1. Model (Q2_K):
  1. Model (Q4_K):

Is this a labeling inconsistency, or could it indicate an issue with the GGUF file format?

The Q2_K file is labeled simply as Q2_K, while the Q4_K file shows Q4_K_M.
Shouldn’t both follow the same naming convention (e.g., Qx_K)?

Could this affect compatibility or is it just a metadata oversight?

1 Like

There are no specific rules for file names…
Either way seems fine.

If it was working fine before (a month ago), then either Ollama or HF was pulling files according to the file name rules, or vice versa.

1 Like

I don’t know, but I have already successfully used q4 model.
As John6666 said, it seems like HF “Use this model” integration became more “strict” and began to check the name in the model metadata with the name of file\quantization.

2 Likes