What is the context length when using ollama to pull HF GGUF

Hi,

I have used Ollama to pull HF GGUF model for local usage according to this post: Use Ollama with any GGUF Model on Hugging Face Hub

The model I have pulled is bartowski/mistralai_Mistral-Small-3.1-24B-Instruct-2503-GGUF · Hugging Face

But I am not sure what is the context length for the pulled model. Is it also as the default as the models from ollama, which is 2048? Or is it the max context length of the model (131072 for this model)

Thanks!

1 Like

It seems that the default is 2048.

By default, Ollama uses a context window size of 2048 tokens.

Hi @John6666 ,

Thanks for the info! Does it mean the context length for all models pulled by Ollama is always 2048 even if the model is from HF?

If that is the case, looks like I have to change the num_ctx manually after pulling the model

1 Like

It seems so.