Hello. I am interested in converting a model to ONNX to get faster inference, but I saw there are two possible approaches:
- Using transformers.onnx package: Export 🤗 Transformers Models
- Using optimum.onnxruntime package: Optimum Inference with ONNX Runtime
Should I convert the model to ONNX with the first and then use it with Optimum? It looks like Optimum can convert models to ONNX by its own now, so what is the point of transformers.onnx package?