Hugging Face Forums
Why transformers doesn't use Multiple GPUs (to increase tokens per second)?
Beginners
John6666
September 21, 2024, 10:59am
2
Possibly:
1 Like
show post in topic
Related topics
Topic
Replies
Views
Activity
Does anyone have an idea how we can run llama2 with multiple GPUs?
🤗Transformers
1
1204
October 26, 2023
Why does Transformer (LLaMa 3.1-8B) give different logits during inference for the same sample when used with single versus multi gpu prediction?
🤗Accelerate
0
46
September 20, 2024
Loading a HF Model in Multiple GPUs and Run Inferences in those GPUs
🤗Accelerate
10
8188
October 16, 2024
Perfectly the same code, single GPU OK, multi GPU ERROR
Beginners
0
26
December 1, 2024
If I use llama 70b and 7b for speculative decoding, how should I put them on my multiple gpus in the code
🤗Transformers
0
18
October 11, 2024