-
Notifications
You must be signed in to change notification settings - Fork 28k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Include "time" as option to save_strategy (and log and eval too!)
Feature request
Request for a new feature
#36310
opened Feb 20, 2025 by
davidhughhenrymack
Transformers does not work well with
with torch.device("meta"):
bug
#36309
opened Feb 20, 2025 by
fxmarty-amd
2 of 4 tasks
Bug about num_update_steps_per_epoch in function _inner_training_loop
bug
#36297
opened Feb 20, 2025 by
onenotell
[Bugs] RuntimeError: No CUDA GPUs are available in transformers v4.48.0 or above when running Ray RLHF example
bug
#36295
opened Feb 20, 2025 by
ArthurinRUC
3 of 4 tasks
[Version: 4.49.0] Qwen2.5-VL is not supported in vLLM because of transformers
bug
#36292
opened Feb 20, 2025 by
usun1997
4 tasks
Cryptic error when using AutoTokenizer with SentencePiece tokenizers without sentencepiece installed
bug
#36291
opened Feb 19, 2025 by
yifanmai
2 of 4 tasks
past_key_value(s) name inconsistency causing problems
bug
#36290
opened Feb 19, 2025 by
HDCharles
2 of 4 tasks
AttributeError: 'dict' object has no attribute 'to_dict'; for Inferencing Lora Merged Qwen/Qwen2.5-VL-3B-Instruct
bug
#36281
opened Feb 19, 2025 by
shahdivax
2 of 4 tasks
The output tensor's data type is not torch.long when the input text is empty.
bug
#36277
opened Feb 19, 2025 by
wangzhen0518
2 of 4 tasks
Torch -> ONNX doesn't work after upgrading transformers to 4.49.0
bug
#36276
opened Feb 19, 2025 by
dongruliu
2 of 4 tasks
Loading a pipeline with
trust_remote_code=True
raises warning
bug
#36273
opened Feb 19, 2025 by
tcapelle
4 tasks
Device Movement Error with 4-bit Quantized LLaMA 3.1 Model Loading
#36272
opened Feb 19, 2025 by
Pritidhrita
3 of 6 tasks
Is T5 model supported with HQQ quantization ? (AttributeError: 'HQQLinear' object has no attribute 'weight')
bug
#36254
opened Feb 18, 2025 by
szwagros
4 tasks
padding_side
is of type bool
when it should be Literal['right', 'left']
bug
#36252
opened Feb 18, 2025 by
winstxnhdw
4 tasks
[BART] Cannot copy out of meta tensor; no data!
bug
#36247
opened Feb 18, 2025 by
jiqing-feng
1 of 4 tasks
AutoModelForCasualLM.from_pretrained()
exits without warning/error
bug
#36245
opened Feb 18, 2025 by
blazgocompany
2 of 4 tasks
A word-level timestamps on whisper generation pipeline is mismatched to total duration
#36228
opened Feb 17, 2025 by
dobby-seo
Incompatibility in flash_attention_2 + Llama + Transformers>=4.43 + Autocast to fp16
bug
#36224
opened Feb 17, 2025 by
poedator
Tensor Parallel performance is not better than eager mode.
bug
#36222
opened Feb 17, 2025 by
jiqing-feng
2 of 4 tasks
Request to add GTrXL: Stabilizing Transformers for Reinforcement Learning
New model
#36220
opened Feb 16, 2025 by
ashok-arora
1 of 2 tasks
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.