hey @haithembrr, your approach sounds very sensible
have you tried playing around with the parameters of the model’s generate
function, e.g. max_len
and trying different strategies like beam search vs sampling (see docs)?
alternatively, you could have a look at the discussion in this thread to see if someone has also run into the same problem: Summarization on long documents