Inference API stopped working

From Hugging Face Discord:

Tom Aarsen

Hello! I believe some of the inference endpoints currently have “Scale to zero” enabled temporarily, meaning they will go down when there’s no usage for a while. The first request will then be slow/fail, but subsequent ones will work. We’re going to remove the scale to zero again so that this is not an issue anymore, apologies for the inconvenience. cc @ VB can you update the scale to zero for the big ST models that already had APIs?

As for my case of my private model, “Scale to zero” was always the case - yes, I needed to wait for 1-3 mins for the first response. But right now it is 404. The issue described in discord is different.

2 Likes

Hi everyone!
Do you have any updates on the fixes for the bugs discussed in this thread?

1 Like

404 with 429 issue is fixed!