The method is described to some extent in the following post. If you want to limit it to the available models, you can add the option Inference=”warm“ .
Hi all ,
I am getting the following error when I call the list_deployed_models() on the InferenceClient.
framework = "text-generation-inference" # "text-to-speech",
deployed_models = client.list_deployed_models([framework])
print(deployed_models)
Error:
BadRequestError: (Request ID: Root=1-67c47933-02bf30da3e80cb5307dc9184;4dbda49a-b0b7-48bc-b0f7-5f7e09fc28d6)
Bad request:
Not allowed to request framework/text-generation-inference for provider hf-inference
Any help is appreciated
Reg…