This is a friendly reminder - the current text generation call will exceed the model’s predefined maximum length (32768). Depending on the model, you may observe exceptions, performance degradation, or nothing at all.
Can anybody justify the this warning while getting the inference in mistralai/Mistral-7B-Instruct-v0.2?