PRO Plan and for running huge models on free inference api?

Hi,

I was curious about whether the pro plan would enable me to do the following:

curl https://api-inference.huggingface.co/models/HuggingFaceH4/starchat-alpha
-X POST
-d ‘{“inputs”: "Can you please let us know more details about your "}’
-H “Authorization: Bearer xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx”

As I am currently getting

{“error”:"The model HuggingFaceH4/starchat-alpha is too large to be loaded automatically (31GB > 10GB)…}

Best,
Jean Elbers

hi @jelber2,

This is a really large model, you may need a dedicated hardware , I recommend you looking at our Inference Endpoints - Hugging Face service and reaching out if you need help, thanks