How do I add a stop token for Inference Endpoints?

How do I add a stop token for Inference Endpoints?

I want to use the Nvidia OpenMath Model and I want to implement stop=["</llm-code>"]