Cannot run large models using API token

I am also unable to run a large model on the inference API, specifically Salesforce/codegen-16B-mono. Neither trying it out using the widget on the website nor using a REST request through python works. In both cases I get a time out, for example the widget gives the following output after some time: Model Salesforce/codegen-16B-mono time out.

Is that because the model is too big or because something in the backend is broken for that model. For the latter case, should I ask the model’s authors for help?