Hugging Face Forums
Loading quantized model on CPU only
๐คTransformers
chanansh
June 1, 2023, 7:19pm
2
I have a similar issue
AssertionError: Torch not compiled with CUDA enabled
show post in topic
Related topics
Topic
Replies
Views
Activity
How to load quantized LLM to CPU only device
Intermediate
0
1876
January 28, 2024
How to parallelize inference on a quantized model
Intermediate
5
192
October 7, 2024
Load quantized model in memory
Beginners
1
580
December 8, 2023
BLOOM models don't run on my GPU using Transformers
๐คTransformers
1
1655
September 18, 2022
How to avert 'loading checkpoint shards'?
๐คTransformers
4
11242
November 1, 2024