Apologies for the confusion, as a Pro user you can access Inference for these special large LLM, read more here as well as higher rate limits for thousands of compatible models on the hub see all tasks here
For custom GPU hardwares and Inference Endpoints follow the pricing here and here
Hugging Face PRO users now have access to exclusive API endpoints for a curated list of powerful models that benefit from ultra-fast inference powered by text-generation-inference. This is a benefit on top of the free inference API, which is available to all Hugging Face users to facilitate testing and prototyping on 200,000+ models. PRO users enjoy higher rate limits on these models, as well as exclusive access to some of the best models available today.
Here are the special list of large language models running with TGI