-
Notifications
You must be signed in to change notification settings - Fork 26.9k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
CUDA Out Of Memory when training a DETR Object detection model with compute_metrics
bug
#34675
opened Nov 9, 2024 by
Kamal-Moha
4 tasks
Vision Encoder-Decoder fails with LLaMA decoder due to missing cross-attention implementation
bug
#34674
opened Nov 9, 2024 by
amazingvince
4 tasks
DataCollatorForMultipleChoice
exists in the docs but not in the package
Feature request
#34671
opened Nov 9, 2024 by
bauwenst
Training wont resuming from checkpoint ( model = Idefics3ForConditionalGeneration.from_pretrained() )
bug
#34660
opened Nov 8, 2024 by
aeltorio
2 of 4 tasks
Neftune computation is probably wrong with packed training
bug
#34659
opened Nov 8, 2024 by
ccdv-ai
4 tasks
Different LlamaRotaryEmbedding in old and new versions of transformers
bug
#34657
opened Nov 8, 2024 by
ivankrylatskoe
4 tasks
Trainer does not call torch.compile when torch_compile=True in TrainingArguments
bug
#34656
opened Nov 8, 2024 by
singularity-s0
2 of 4 tasks
The support of Request for a new feature
Mllama
in AutoModel
Feature request
#34654
opened Nov 8, 2024 by
HanGyeol-Yoo
Add functionality for deleting adapter layers in PEFT integration
Feature request
Request for a new feature
#34649
opened Nov 8, 2024 by
itsskofficial
Passing nn.Parameter values within the model architecture as deep copies.
bug
#34643
opened Nov 7, 2024 by
James6Chou
2 of 4 tasks
Inconsistent Hyperparameter Application in Distributed DeepSpeed + Optuna Setup for Hyperparameter Search
bug
#34641
opened Nov 7, 2024 by
corentin-ryr
4 tasks
Different loss when using different "attn_implemetation", when using "flash_attention_2" is notebly larger than "eager" and "spda".
bug
#34640
opened Nov 7, 2024 by
ttw1018
2 of 4 tasks
Padding error when using Universal Assisted Generation with ASR pipeline
bug
#34639
opened Nov 7, 2024 by
danielkorat
4 tasks
When will transformers trainer supports fp8
Feature request
Request for a new feature
#34635
opened Nov 7, 2024 by
wangxiaoyang-dev
safe_globals are needed to resume training on upcoming PyTorch 2.6
#34631
opened Nov 6, 2024 by
dvrogozh
Issue with SynthID watermark implementation
Feature request
Request for a new feature
#34630
opened Nov 6, 2024 by
marluxiaboss
IdeficsProcessor has wrong __call__ args and the docstring is not updated
bug
Documentation
Vision
#34628
opened Nov 6, 2024 by
Kami-chanw
2 of 4 tasks
[openai/whisper-tiny][torch.compile] Model compilation: AttributeError: 'DynamicCache' object has no attribute 'key_cache'
Audio
bug
#34626
opened Nov 6, 2024 by
daniil-lyakhov
2 of 4 tasks
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.