How do I add a stop token for Inference Endpoints?
I want to use the Nvidia OpenMath Model and I want to implement stop=["</llm-code>"]
How do I add a stop token for Inference Endpoints?
I want to use the Nvidia OpenMath Model and I want to implement stop=["</llm-code>"]