Adapting a model from Spaces to Inference Endpoint

Hello @plexus3d,

You can directly deploy your custom inference pipeline as an inference endpoint using a custom container. This would mean creating a model repository with your space code and then using a custom docker with gradio. Here is an example: philschmid/space-naver-donut-cord · Hugging Face.

Additionally, as you mentioned, you can create a custom inference handler. In the documentation, we have several examples on how to do this: * Optimum and ONNX Runtime


Or you can use the API from your Spaces