Hugging Face Forums
Beam_search bottlenecks inference with only 1 used cpu
🤗Transformers
adelplace
October 13, 2022, 12:06pm
2
It seems like I am not the only one facing this problem :
Any ideas of solution ?
show post in topic
Related topics
Topic
Replies
Views
Activity
Is there any way to avoid CPU bottlenecks when doing single prompt inference?
Intermediate
1
947
June 12, 2023
Multiple gpu not properly parallelized during model.generate()
🤗Transformers
4
1581
October 9, 2022
Very low GPU usage when translating text, datasets not helping
🤗Transformers
3
5667
July 12, 2022
NLP Pretrained model model doesn’t use GPU when making inference
🤗Transformers
11
10020
March 11, 2022
Model.generate() is extremely slow while using beam search
🤗Transformers
2
5268
July 24, 2022