Query endpoint for LLM

For a personal project Im trying to summarize news articles.
I have locally launched and played with the following model = “facebook/bart-large-cnn” and it works great.
However I wanted to know whether there was a free endpoint I could use for that model?
Ideally I would send it requests, it’d run on HF servers and then send back the query answer.
This is quite similar to what we can do with the OpenAI GPT api except the latter is pretty expensive.
If not, can you point me towards a free endpoint?