Prompt caching in pipelines

I have a very long system prompt and a large dataset of short queries appended to the prompt. I would like to use caching mechanism so I don’t process the long prompt again each time.
I have read how to cache common instruction prompt and re-use cache to continue generation but I would like to use this caching mechanism together with pipelines.
More precisely I have a text-generation pipeline for which I would like to provide a system prompt which would be reused for each generation.
Is there such a pipeline available?
If not, how could I make my own text-generation pipeline that mimics the base class but uses the caching mechanism?

1 Like

Hmm… Perhaps not yet available…?