Hugging Face Forums
Loading quantized model on CPU only
๐คTransformers
chanansh
June 1, 2023, 7:19pm
2
I have a similar issue
AssertionError: Torch not compiled with CUDA enabled
show post in topic
Related topics
Topic
Replies
Views
Activity
How to load quantized LLM to CPU only device
Intermediate
0
1970
January 28, 2024
Load quantized model in memory
Beginners
1
599
December 8, 2023
Does load_in_8bit directly load the model in 8bit? (spoliler, do not seem like it)
Beginners
0
1502
July 11, 2023
Fine-tuning with load_in_8bit and inference without load_in_8bit possible?
๐คTransformers
4
24439
August 23, 2022
Want to use CPU for falcon7b
Beginners
0
319
June 22, 2023