-
Notifications
You must be signed in to change notification settings - Fork 28.8k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Failed to load santacoder model with transformer 4.51.3, it's a similar issue like #37737
bug
#37765
opened Apr 24, 2025 by
nv-guomingz
8 tasks
[Install] fail to install pip install transformers==4.42
bug
#37752
opened Apr 24, 2025 by
vfdff
4 tasks
Performance degradation on certain vision models from v4.51.*
bug
#37748
opened Apr 24, 2025 by
yuan-thomas
3 of 4 tasks
Failed to load model with transformers 4.51.3 when WORLD_SIZE set to 1 on nvidia gpu
bug
#37737
opened Apr 24, 2025 by
nv-guomingz
4 tasks
ValueError: GGUF model with architecture deci is not supported yet.
Feature request
Request for a new feature
#37736
opened Apr 24, 2025 by
iEddie-cmd
tokenizer
is still being used in Trainer
instead of processing_class
#37734
opened Apr 24, 2025 by
arjunaskykok
Facing "NotImplementedError:
config.do_stable_layer_norm is False
is currently not supported" in FlaxWav2vec2Models
bug
Flax
#37731
opened Apr 24, 2025 by
premmurugan229
2 of 4 tasks
Gemma3 is ExecuTorch compatible
Feature request
Request for a new feature
#37727
opened Apr 23, 2025 by
guangy10
Make Request for a new feature
argmax
in post_process_semantic_segmentation
optional
Feature request
#37715
opened Apr 23, 2025 by
simonreise
Loading and Saving Pretrained model to the same directory raises SafeTensorError: IOError
bug
#37713
opened Apr 23, 2025 by
minerharry
3 of 4 tasks
Can't perform inference with images on Gemma-3-12b-it-qat-int4.0
bug
#37710
opened Apr 23, 2025 by
njemanzedavid
2 of 4 tasks
last_cache_position
definition issue in hybrid SWA models
bug
#37706
opened Apr 23, 2025 by
plienhar
[i18n-Chinese] Translating model_doc/bert.md to Chinese
WIP
Label your PR/Issue with WIP for some long outstanding Issues/PRs that are work in progress
#37705
opened Apr 23, 2025 by
Nanji-Huaji
Tokenizing with
apply_chat_template
behaves differently from regular tokenizing
bug
#37686
opened Apr 22, 2025 by
sayanshaw24
2 of 4 tasks
Behaviour of
batch_eval_metrics
determines the include_for_metrics
behaviour
bug
#37683
opened Apr 22, 2025 by
prabhuteja12
1 of 4 tasks
Model.from_pretrained
breaks when using SinusoidalEmbedding
bug
#37671
opened Apr 22, 2025 by
ZhiyuanChen
4 tasks
Avoid adding space when decoding tokenization
Feature request
Request for a new feature
#37659
opened Apr 21, 2025 by
cikay
Adding Paged Attention to Qwen1.5-MoE-A2.7B-Chat models using PyTorch XLA and Pallas
Feature request
Request for a new feature
#37648
opened Apr 21, 2025 by
ranwangmath1988
"pipeline" is not exported from module "transformers"
bug
#37646
opened Apr 21, 2025 by
tekumara
1 of 4 tasks
Error message is misleading for missing protobuf
bug
#37641
opened Apr 20, 2025 by
Ishan-Kumar2
2 of 4 tasks
Processor multiprocessing error when load custom processor
bug
#37637
opened Apr 20, 2025 by
Kuangdd01
2 of 4 tasks
Previous Next
ProTip!
Updated in the last three days: updated:>2025-04-21.