Raise Inference Client GB Limit

Hello, is there by any chance a way to increase the original 10 GB limit placed on Inference Client for enterprises on HF? Thanks.

Hi @ThatOneCoder Currently there is a 10GB limit placed on the Inference Client.

The free Inference API is a solution to easily explore and evaluate models, and Inference Endpoints is our paid inference solution for production use cases. For larger models or volumes of requests, or if you need guaranteed latency/performance, we recommend using Inference Endpoints instead to easily deploy your models on dedicated, fully-managed infrastructure.

Further pricing information can be found here.

Okay, thanks.

This topic was automatically closed 12 hours after the last reply. New replies are no longer allowed.