It has 192GB of VRAM!
Ollama’s default is Q4_K_M (if you don’t specify otherwise, it uses this, which is unlikely to cause problems), so I think it will be more than enough. The amount of consumption is a little over 64GB for the model, plus a little for inference. If there are a lot of people using it at the same time, or if you want to process very long sentences, it will use more VRAM, but it will still be unlikely to cause problems.
By the way, Ollama is also fast enough, but Llamacpp seems to be even faster. If you’re having trouble with speed, you might want to try changing it.