PRO Plan and for running huge models on free inference api?


I was curious about whether the pro plan would enable me to do the following:

-d ‘{“inputs”: "Can you please let us know more details about your "}’
-H “Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx”

As I am currently getting

{“error”:"The model HuggingFaceH4/starchat-alpha is too large to be loaded automatically (31GB > 10GB)…}

Jean Elbers

hi @jelber2,

This is a really large model, you may need a dedicated hardware , I recommend you looking at our Inference Endpoints - Hugging Face service and reaching out if you need help, thanks