'You have passed more than 3000 mel input features (> 30 seconds) which automatically enables long-form generation which requires the model to predict timestamp tokens. Please either pass `return_timestamps=True` or make sure to pass no more than 3000 mel input features.', 'warnings': ['There was an inference error: You have passed more than 3000 mel input features (> 30 seconds) which automatically enables long-form generation which requires the model to predict timestamp tokens. Please either pass `return_timestamps=True` or make sure to pass no more than 3000 mel input features.
also using 1, True (boolean), “True” But the error still comes up. I am also able to see the error on the UI playground at openai/whisper-large-v3-turbo · Hugging Face : Did anyone see similar behaviour ? Can the team please help on this ? Many thanks
I also faced the same problem (whisper-large-v3-turbo, ogg audio). Until yesterday it was working fine, but now audio >30s causing error, return_timestamps does not help as well. Same issue on whisper-large-v3
I agree with @TOOTHED - seeing same behaviour here. Files >30 s used to work without any issues but now throw an error, and no keyword/param works, should this be raised with the API team somehow ?
I think we can contact the Hugging Face team from any of these. The Hugging Face functions are managed on github with the exception of a few, so raising an issue on github is the most reliable way to get in touch. However, there are so many github pages that it’s hard to know where to write…
Hugging Face Documents and Hub general issue
Library handling Inference API
Whisper Class (Maybe it’s different this time…)
Well, tbh it is literally my first comment on hf forums, so I am not sure how it is working here. But overall, yes, as whisper models were not updated and this behaviour is similar for all them, the error is for sure caused by the updates in hugging face API - either AutomaticSpeechRecognition method (i didn’t test if any parameters work), or something deeper
I have now cross-posted this in three different places - wondering if someone knows how to tag someone from the dev/support team. I feel like this is part of the broader inference api changes, the behaviour of this whisper endpoint has radically changed and is barely usable - the silence from the team on all open issues here is somewhat worrying.
If it’s related to Whisper, Serverless Inference API, or servers, then michellehbn, Wauplin, pierric, or victor?
You could also ask meganariley…
The last resort is julien-c.