Skip to content
This repository was archived by the owner on Oct 9, 2024. It is now read-only.

Issues: huggingface/transformers-bloom-inference

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

does this work for llama 65B
#98 by GradientGuru was closed Jul 31, 2023
accelerate in bloom-inference-scripts?
#91 by jeromeku was closed May 23, 2023
Unable to reload a quantized model
#85 by moonlightian was closed May 10, 2023
The details of hf-accelerate pp.
#83 by tohneecao was closed May 20, 2023
BUILD ERROR with nvcc
#81 by tohneecao was closed May 10, 2023
tokenizer.json 乱码怎么解析
#78 by hongshengxin was closed May 10, 2023
concurrent requests
#75 by ustclan was closed Apr 7, 2023
beam search
#73 by syp1997 was closed Apr 19, 2023
Short response for bloom inferring
#70 by raihan0824 was closed Mar 29, 2023
Should I use bf16 or fp16?
#69 by richarddwang was closed Mar 23, 2023
how to run the server?
#64 by raihan0824 was closed Mar 29, 2023
OOM of CUDA when using one GPU
#60 by xiongjun19 was closed May 31, 2023
ProTip! no:milestone will show everything without a milestone.