Quantized Model size difference when using Optimum vs. Onnxruntime

I see you already opened issue at optimum: It makes sense to focus the discussion on one platform.

Feel free to either close the issue or this thread.