How can we fine-tune an Integer Bert for question answering task?
As explained in IBERT’s model card: fine-tuning the model consists of 3 stages:
- full precision fine-tuning
- quantization-aware training.
Step 2 is simply setting the
quantize attribute of the model’s configuration to
Step 3 is the same as step 1, but now with your quantized model.