I was trying to make an inference container using optimum and onnxruntime-gpu. It works well in my machine (T4), but I face this error when I do it inside a docker container:
Asked to use CUDAExecutionProvider as an ONNX Runtime execution provider, but the available execution providers are ['AzureExecutionProvider', 'CPUExecutionProvider'].
I am using this container as a starting point: nvidia/cuda:12.2.2-base-ubuntu22.04
Also I don’t see docker images for optimum with cuda 12.x on dockerhub: Docker