Hugging Face Forums
Deepspeed inference and infinity offload with bitsandbytes 4bit loaded models
🤗Transformers
DeepSpeed
AvivB
July 27, 2023, 8:46pm
3
I am getting this error too.
Edit: There appears to be support now:
show post in topic
Related topics
Topic
Replies
Views
Activity
Finetuning LLama2-70B using 4-bit quantization on multi-GPU using Deepspeed ZeRO
Intermediate
1
2443
March 19, 2024
Deepspeed inference stage 3 + quantization
DeepSpeed
0
1014
March 8, 2024
Is Int8 quantization training possible while using deepspeed?
DeepSpeed
0
587
December 1, 2021
Deepspeed ZeRO2, PEFT, bitsnbytes training
DeepSpeed
0
129
June 4, 2024
An error i ve been trying to fix for days now
Intermediate
4
495
November 19, 2024