-
Notifications
You must be signed in to change notification settings - Fork 28k
Pull requests: huggingface/transformers
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[bugfix] Update modeling_llama.py so it skips keys correctly
#36289
opened Feb 19, 2025 by
HDCharles
Loading…
[modular] allow multiple modular files in the same model folder
#36287
opened Feb 19, 2025 by
Cyrilvallez
Loading…
[WIP] Set default processing device to best available device in fast image processors
#36268
opened Feb 18, 2025 by
yonigozlan
Loading…
Remove hardcoded slow image processor class in processors supporting fast ones
#36266
opened Feb 18, 2025 by
yonigozlan
Loading…
Upgrading torch version and cuda version in quantization docker
#36264
opened Feb 18, 2025 by
MekkCyber
Loading…
Add D-FINE Model into Transformers
#36261
opened Feb 18, 2025 by
VladOS95-cyber
Loading…
4 of 5 tasks
[generate] remove legacy cache in
t5
and whisper
-based models (deprecated in v4.48)
#36238
opened Feb 17, 2025 by
gante
Loading…
Prevent Reinitialization of Resized LM Head When
tie_word_embeddings
is False #35141
#36221
opened Feb 16, 2025 by
sambhavnoobcoder
Loading…
fix: prevent second save in the end of training if last step was saved already
#36219
opened Feb 16, 2025 by
NosimusAI
Loading…
2 of 5 tasks
Improvements in attention_forward functions
#36218
opened Feb 15, 2025 by
mseeger
Loading…
3 of 5 tasks
[WIP] Add a dedicated tokenizer for byte level transformers
#36216
opened Feb 15, 2025 by
apehex
Loading…
Previous Next
ProTip!
Follow long discussions with comments:>50.