I ran into a similar issue, the data in your jsonl file for inference should look something like this…
{"inputs":"...Some long text string, likely over 512 tokens after tokenization....","parameters":{"truncation":true,"max_length":512}}
{"inputs":"...Another long text string, likely over 512 tokens after tokenization....","parameters":{"truncation":true,"max_length":512}}