-
Notifications
You must be signed in to change notification settings - Fork 27.9k
Pull requests: huggingface/transformers
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Fix potential regex catastrophic backtracking in NougatTokenizerFast
#36121
opened Feb 10, 2025 by
Rocketknight1
•
Draft
[generate] revert change in Aria: the maximum cache length must match
max_length
#36120
opened Feb 10, 2025 by
gante
Loading…
Proper performant flex attention implementation
#36103
opened Feb 8, 2025 by
bursteratom
Loading…
3 of 15 tasks
Remove loading custom kernel for RT-DETRv2
cleanup
Vision
#36098
opened Feb 7, 2025 by
qubvel
Loading…
Add pipeline parallel plan to
PretrainedConfig
and PreTrainedModel
#36091
opened Feb 7, 2025 by
hmellor
Loading…
5 tasks
Fix: Llama - adjust the rotary embedding dimensions.
#36090
opened Feb 7, 2025 by
abdullahselek
Loading…
1 of 5 tasks
qwen2.5vl: fix bugs when using flash2+bf16 or num_return_sequences>1
#36083
opened Feb 7, 2025 by
gewenbin0992
Loading…
5 tasks
PyTorch's int8 & int4 WoQ GEMMs expect contiguous activations
#36078
opened Feb 7, 2025 by
sanchitintel
Loading…
add special tokens should probably be serialized and deserialized?
#36070
opened Feb 6, 2025 by
ArthurZucker
•
Draft
Fix bug in prepare_inputs_for_generation function: in Qwen2-VL (#36037)
#36038
opened Feb 4, 2025 by
JamesHujy
Loading…
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.