Optimize large scale transformer model inference with ONNX Runtime

Hello,

I would like to ask if the following transformer models could be optimized through ONNX:

  • ‘joeddav/xlm-roberta-large-xnli’
  • ‘t5-base’
  • ‘sentence-transformers/all-roberta-large-v1’

Thank you in advance!

Mariya