Optimize AND quantize with Optimum

I managed to make the example from the blog Accelerated Inference with Optimum and Transformers Pipelines
work with the advise from this thread, but since this week the code breaks when quantizing the model
with this error

RuntimeError: Unable to find data type for weight_name='/roberta/encoder/layer.0/attention/output/dense/MatMul_output_0'

I have created a new post to explain the steps how to reproduce the error Optimum library optimization and quantization fails - #2 by ddahlmeier