Thanks for the comment, but that still doesnât solve the issue. I tried again, same error message. Iâm not sure (and doubt) if 4-bit model is still not supported? If so, what am I doing wrong?
Can someone please provide answer or insight to this issue?
The docs arenât entirely clear, but my read is that 8-bit is possible but 4-bit is not:
Note that once a model has been loaded in 4-bit it is currently not possible to push the quantized weights on the Hub. Note also that you cannot train 4-bit weights as this is not supported yet. However you can use 4-bit models to train extra parameters, this will be covered in the next section.
Yeah, correct, with the caveat that you still need to install transformers from source as the latest stable release doesnât support it. Hence why I tagged the issue above, cheers.