Skip to content
This repository was archived by the owner on Oct 9, 2024. It is now read-only.

Issues: huggingface/transformers-bloom-inference

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

DeepSpeed runtime partition failed
#18 by lanking520 was closed Oct 11, 2022
transformers_bloom_parallel link 404
#2 by zcrypt0 was closed Sep 19, 2022
Inference returns nan log-probability
#38 by vinhngx was closed Feb 7, 2023
OOM of CUDA when using one GPU
#60 by xiongjun19 was closed May 31, 2023
how to run the server?
#64 by raihan0824 was closed Mar 29, 2023
does this work for llama 65B
#98 by GradientGuru was closed Jul 31, 2023
Inference hangs after GPU OOM
#32 by xiang-deng was closed Nov 22, 2022
ProTip! Add no:assignee to see everything that’s not assigned.