Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Serialization error when tokenizer_config key matches function name in PreTrainedTokenizerBase
#30796
opened May 14, 2024 by
avnermay
TFSequenceClassificationLoss for MultiLabel classification
#30792
opened May 13, 2024 by
ds-mike
2 of 4 tasks
Can the BNB quantization process be on GPU?
Quantization
#30770
opened May 13, 2024 by
mxjmtxrm
4 tasks
ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided []
trainer
#30769
opened May 12, 2024 by
gtanya89
4 tasks
Issues occuring during parallel evaluation (using Trainer.evaluate)
trainer
#30767
opened May 12, 2024 by
psychocosine
4 tasks done
For multiple GPUs: torch.cuda.empty_cache() stuck forever
trainer
#30766
opened May 11, 2024 by
animeshkumarpaul
2 of 4 tasks
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained
#30762
opened May 11, 2024 by
yingqianch
4 tasks
BART generate with min_new_tokens exceeds maximum length
Generation
#30759
opened May 11, 2024 by
vsocrates
2 of 4 tasks
Implement kv cache sparsity like H2O with attention score
Cache
Feature request
Request for a new feature
#30758
opened May 11, 2024 by
HarryWu99
TokenClassificationPipeline support is_split_into_words tokeniser parameter
Core: Pipeline
Internals of the library; Pipeline.
#30757
opened May 11, 2024 by
swtb3
TypeError: 'list' object is not callable || Resume from checkpoint
PEFT
#30754
opened May 11, 2024 by
satpalsr
4 tasks
recent version of Transformers seems to mess with forward/__call__. Breaks patching loss function
#30753
opened May 10, 2024 by
grahamannett
3 of 4 tasks
train_new_from_iterator does not properly modify the tokenizer's postprocessor's ids when using a Sequence postprocessor
#30752
opened May 10, 2024 by
dmcinerney
1 of 4 tasks
BitsNBytes 4 bit quantization error message typo and logical errors in error message handling
Quantization
#30751
opened May 10, 2024 by
jkterry1
4 tasks
use_reentrant=False can't be set properly
DeepSpeed
trainer
#30749
opened May 10, 2024 by
getao
2 of 4 tasks
[Batched Whisper] ValueError on input mel features
Audio
#30740
opened May 10, 2024 by
kerem0comert
2 of 4 tasks
[DOCS] - Model outputs of RecurrentGemmaCausalLM doesn't align with the documentation
#30736
opened May 10, 2024 by
godjw
4 tasks
Meet problems when I use the file src/transformers/models/llama/convert_llama_weights_to_hf.py to transfer LlaMa-7B
#30734
opened May 9, 2024 by
wwxxyy1996
2 of 4 tasks
Mixtral past_key_values and output_router_logits incompatible
#30731
opened May 9, 2024 by
sorgfresser
2 of 4 tasks
Support for Multiple Datasets and Domain-Specific Loss Calculation in Trainer
Feature request
Request for a new feature
trainer
#30725
opened May 9, 2024 by
Ajmalshamsudheen
hub_strategy="every_save"
won't push the model to the Hub if large
#30724
opened May 9, 2024 by
alvarobartt
2 of 4 tasks
Previous Next
ProTip!
Adding no:label will show everything without a label.