Fail: [ONNXRuntimeError] : 1 : FAIL : Deserialize tensor onnx:

Hi @serdarcaglar , currently this is an issue with models with external data format. We have a PR [255] open for the issue and the fix should be available soon :smile: . In the mean time you could run the above model by disabling the cache.

from datasets import load_dataset
from transformers import AutoProcessor, pipeline
from optimum.onnxruntime import ORTModelForSpeechSeq2Seq

 processor = AutoProcessor.from_pretrained("openai/whisper-large")
 model = ORTModelForSpeechSeq2Seq.from_pretrained("openai/whisper-large", from_transformers=True, use_cache=False)
 speech_recognition_pipeline = pipeline(
    "automatic-speech-recognition",
     model=model,
     feature_extractor=processor.feature_extractor,
     tokenizer=processor.tokenizer,
  )
1 Like