Hugging Face Forums
Why transformers doesn't use Multiple GPUs (to increase tokens per second)?
Beginners
John6666
September 21, 2024, 10:59am
2
Possibly:
1 Like
show post in topic
Related topics
Topic
Replies
Views
Activity
How to generate with a single gpu when a model is loaded onto multiple gpus?
Beginners
0
889
February 9, 2024
Getting error when running inference in multiple GPUs
🤗Transformers
0
656
October 13, 2023
[SOLVED] What's the right way to do GPU paralellism for inference (not training) on AutoModelForCausalLM?
🤗Transformers
1
237
August 26, 2024
If I use llama 70b and 7b for speculative decoding, how should I put them on my multiple gpus in the code
🤗Transformers
0
48
October 11, 2024
Using 3 GPUs for training with Trainer() of transformers
🤗Transformers
2
2323
October 18, 2023