-
Notifications
You must be signed in to change notification settings - Fork 4.9k
Issues: hiyouga/LLaMA-Factory
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
GRPO 在搞了吗?
bug
Something isn't working
pending
This problem is yet to be addressed
#6875
opened Feb 9, 2025 by
Harryjun
1 task done
ultra_chat数据集使用问题
bug
Something isn't working
pending
This problem is yet to be addressed
#6871
opened Feb 9, 2025 by
Fangkang515
1 task done
Support for finetuning the convolutional layers of VLMs with lora
enhancement
New feature or request
pending
This problem is yet to be addressed
#6869
opened Feb 8, 2025 by
roger-creus
1 task done
Docker部署后使用Nginx代理访问,部分接口与资源请求宿主机IP与端口形式访问
bug
Something isn't working
pending
This problem is yet to be addressed
#6863
opened Feb 8, 2025 by
lanyur
1 task done
Qwen2VL Video features and video tokens do not match error
bug
Something isn't working
pending
This problem is yet to be addressed
#6859
opened Feb 8, 2025 by
BubblyYi
1 task done
UnicodeDecodeError: 'utf-8' codec can't decode byte 0xce in position 0: invalid continuation byte
bug
Something isn't working
pending
This problem is yet to be addressed
#6856
opened Feb 8, 2025 by
MengHao666
1 task done
bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cget_managed_ptr
bug
Something isn't working
pending
This problem is yet to be addressed
#6851
opened Feb 7, 2025 by
missTL
1 task done
NCCL broken pipe
bug
Something isn't working
pending
This problem is yet to be addressed
#6849
opened Feb 7, 2025 by
Eggwardhan
1 task done
使用vllm部署DeepSeek-R1-Distill-Qwen-7B模型时,显存翻倍
bug
Something isn't working
pending
This problem is yet to be addressed
#6846
opened Feb 7, 2025 by
nvliajia
1 task done
Qwen2.5-VL-7B Instruct Lora model vllm load failed
bug
Something isn't working
pending
This problem is yet to be addressed
#6845
opened Feb 7, 2025 by
shenhaitao010
1 task done
Use multipage and sidebar components in the web app
enhancement
New feature or request
pending
This problem is yet to be addressed
#6841
opened Feb 6, 2025 by
yvrjsharma
1 task done
llamafactory最新版0.9.2.dev0,unsloth加速训练报错
bug
Something isn't working
pending
This problem is yet to be addressed
#6836
opened Feb 6, 2025 by
yecphaha
1 task done
qwen2.5-72B-Instruct训练一半数据卡住
bug
Something isn't working
pending
This problem is yet to be addressed
#6832
opened Feb 6, 2025 by
CuiXinYu123
1 task done
CUDA OOM in the middle of QLoRA with Llama 3.3 70B 4-bit AWQ
bug
Something isn't working
pending
This problem is yet to be addressed
#6827
opened Feb 5, 2025 by
paolovic
1 task done
求助:如何对 DeepSeek R1 进行 SFT
bug
Something isn't working
pending
This problem is yet to be addressed
#6824
opened Feb 5, 2025 by
yuchunyu97
1 task done
504 Gateway Time-out
bug
Something isn't working
pending
This problem is yet to be addressed
#6822
opened Feb 5, 2025 by
MarkJiang-maji
1 task done
全量训练MiniCPM-o问题
bug
Something isn't working
pending
This problem is yet to be addressed
#6819
opened Feb 5, 2025 by
JACKYLUO1991
1 task done
NPU ds3_ofld训练不释放内存最终OOM
bug
Something isn't working
npu
This problem is related to NPU devices
pending
This problem is yet to be addressed
#6816
opened Feb 5, 2025 by
ultramangod
1 task done
Do you support Mamba-Codestral-7B?
enhancement
New feature or request
pending
This problem is yet to be addressed
#6808
opened Feb 4, 2025 by
displaywz
1 task done
lora微调minicpm-o-2.6 卡着不动,A6000 48G * 8卡,batchsize = 1,但是单卡又可以微调,拉的最新的代码,看issuse很多出现这个问题但都没解决,qwen2vl 多卡微调没有卡死
bug
Something isn't working
pending
This problem is yet to be addressed
#6806
opened Feb 4, 2025 by
cqray1990
1 task done
结合 GRPO 支持 DeepSeek-R1 等推理模型的复现,达到 huggingface open-r1 的类似效果
enhancement
New feature or request
pending
This problem is yet to be addressed
#6792
opened Feb 2, 2025 by
submartingales
1 task done
Qwen2.5-VL full sft dtype error
bug
Something isn't working
pending
This problem is yet to be addressed
#6791
opened Feb 2, 2025 by
wyuc
1 task done
ValueError: The checkpoint you are trying to load has model type Something isn't working
pending
This problem is yet to be addressed
llava_mistral
but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
bug
#6781
opened Jan 30, 2025 by
dainini
1 task done
有计划支持Deepseek的janus pro微调么
enhancement
New feature or request
pending
This problem is yet to be addressed
#6775
opened Jan 28, 2025 by
mkygogo
1 task done
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.