Skip to content

Pull requests: huggingface/transformers

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

fix: typos in documentation files
#36122 opened Feb 10, 2025 by maximevtush Loading…
update awesome-transformers.md.
#36115 opened Feb 10, 2025 by zhanluxianshen Loading…
2 of 5 tasks
add DeepSpeed tensor parallel initialization.
#36114 opened Feb 10, 2025 by inkcherry Loading…
Proper performant flex attention implementation
#36103 opened Feb 8, 2025 by bursteratom Loading…
3 of 15 tasks
Fixup another model + encoder decoder
#36095 opened Feb 7, 2025 by muellerzr Draft
5 tasks
Fix: Llama - adjust the rotary embedding dimensions.
#36090 opened Feb 7, 2025 by abdullahselek Loading…
1 of 5 tasks
fix: dtype might change during resize
#36089 opened Feb 7, 2025 by LarsHaalck Loading…
2 of 5 tasks
[docs] fix bug in deepspeed config
#36081 opened Feb 7, 2025 by faaany Loading…
[docs] update awq doc
#36079 opened Feb 7, 2025 by faaany Loading…
[Bugfix] Fix reloading of pixtral/llava configs
#36077 opened Feb 7, 2025 by kylesayrs Loading…
Speedup modular conversion w/ multiproc
#36073 opened Feb 6, 2025 by muellerzr Draft
1 of 5 tasks
Optim: APOLLO optimizer integration
#36062 opened Feb 6, 2025 by zhuhanqing Loading…
🚧 [WiP] Add Janus model
#36053 opened Feb 5, 2025 by yaswanth19 Draft
5 tasks
Remove type hint Unpack[FlashAttentionKwargs]
#36049 opened Feb 5, 2025 by ydshieh Loading…
Anole add model
#36047 opened Feb 5, 2025 by zucchini-nlp Loading…
Add Phi-3.5-vision
#36036 opened Feb 4, 2025 by Dahlbomii Draft
[core] Large/full refactor of from_pretrained
#36033 opened Feb 4, 2025 by Cyrilvallez Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.