Hugging Face Forums
Beam_search bottlenecks inference with only 1 used cpu
🤗Transformers
adelplace
October 13, 2022, 12:06pm
2
It seems like I am not the only one facing this problem :
Any ideas of solution ?
show post in topic
Related topics
Topic
Replies
Views
Activity
Is there any way to avoid CPU bottlenecks when doing single prompt inference?
Intermediate
1
965
June 12, 2023
Multiple gpu not properly parallelized during model.generate()
🤗Transformers
4
1609
October 9, 2022
Very low GPU usage when translating text, datasets not helping
🤗Transformers
3
5783
July 12, 2022
NLP Pretrained model model doesn’t use GPU when making inference
🤗Transformers
11
10082
March 11, 2022
Model.generate() is extremely slow while using beam search
🤗Transformers
2
5340
July 24, 2022