-
Notifications
You must be signed in to change notification settings - Fork 27.4k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
ModernBERT inference fails on CPU: ValueError: Pointer argument (at 0) cannot be accessed from Triton (cpu tensor?)
bug
#35388
opened Dec 21, 2024 by
umarbutler
4 tasks
Potentially incorrect calculation of
total_updates
on >=4.46.0 since #34198 affecting multi gpu training
bug
#35387
opened Dec 21, 2024 by
chiragjn
2 of 4 tasks
modernbert logits do not have gradient
bug
#35386
opened Dec 21, 2024 by
andersonbcdefg
3 of 4 tasks
Support modernBERT for encoder-decoder models
Feature request
Request for a new feature
#35385
opened Dec 21, 2024 by
Bachstelze
MultiModalityCausalLM does not support Flash Attention 2.0 yet
#35383
opened Dec 21, 2024 by
AlanPonnachan
RuntimeError: self and mat2 must have the same dtype, but got Float and BFloat16
when training with torch_compile
bug
#35382
opened Dec 21, 2024 by
umarbutler
2 of 4 tasks
is_causal arg appears twice in FAttention call from GPT2Attention.forward()
bug
#35380
opened Dec 21, 2024 by
poedator
2 of 4 tasks
'do_sample' model default cannot be overridden
bug
#35372
opened Dec 20, 2024 by
Zoher15
2 of 4 tasks
Model loaded with
PretrainedModel.from_pretrained
and with torch.device("cuda"):
decorator leads to unexpected errors compared to .to("cuda")
bug
#35371
opened Dec 20, 2024 by
fxmarty-amd
2 of 4 tasks
Default value for mean_resizing in resize_token_embeddings should be False
bug
#35357
opened Dec 20, 2024 by
cyr0930
4 tasks
Maybe the way SequenceClassification Model calculates the last non-pad token is not reasonable.
bug
#35352
opened Dec 20, 2024 by
liangxuZhang
4 tasks
SinkCache (StreamLLM) implemented over Post-RoPE Key cache might result in confused position for inference
bug
#35350
opened Dec 19, 2024 by
wangguangtao0722
4 tasks
A warning message showing that
MultiScaleDeformableAttention.so
is not found in /root/.cache/torch_extensions
if ninja
is installed with transformers
bug
#35349
opened Dec 19, 2024 by
cainmagi
1 of 4 tasks
[Request for a new feature
Mamba2
] Varlen implementation
Feature request
#35346
opened Dec 19, 2024 by
vasqu
Llama model, torch.compile output for custom device does not match with eager/cpu when generation_config.use_cache set to True
bug
#35343
opened Dec 19, 2024 by
vpandya-quic
4 tasks
Option to Disable Model Caching When Using "pipeline"
Feature request
Request for a new feature
#35337
opened Dec 19, 2024 by
FadiAmon
Default arguments in
DebertaConfig
disable relative attention, contrary to the docs and deberta-base
bug
#35335
opened Dec 19, 2024 by
bauwenst
4 tasks
DeBERTa's
DisentangledSelfAttention
hardcodes float
dtype, which causes bfloat16
overflow error
bug
#35332
opened Dec 19, 2024 by
bauwenst
2 of 4 tasks
tokenizer decode decode with timestamp fails for extended vocabulary
bug
#35330
opened Dec 18, 2024 by
bnestor
2 of 4 tasks
InternVL is ExecuTorch Compatible
ExecuTorch
Feature request
Request for a new feature
#35327
opened Dec 18, 2024 by
guangy10
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.