This repository was archived by the owner on Oct 9, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 112
Issues: huggingface/transformers-bloom-inference
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
8x40GB GPU instance failing with microsoft/bloom-deepspeed-inference-int8
#8
by tjarmain
was closed Feb 18, 2023
running error with "Bus error: nonexistent physical address"
#16
by Emerald01
was closed Feb 18, 2023
[Question] 4-GPU shard
microsoft/bloom-deepspeed-inference-int8
#4
by zcrypt0
was closed Sep 16, 2022
Newly added model_class argument has not been reflected to server code
#33
by koreyou
was closed Nov 16, 2022
RuntimeError: Error building extension 'transformer_inference'
#48
by Mahyar-Ali
was closed Feb 2, 2023
Max tokens generated remains constant for whatever the input token size
#55
by vamsikrishnav
was closed Feb 21, 2023
Why is the throughput of DS-inference doubled when using 4 A100 GPUs compared to 8 A100 GPUs
#59
by DominickZhang
was closed Apr 6, 2023
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.