-
Notifications
You must be signed in to change notification settings - Fork 117
Issues: thu-ml/SageAttention
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
RuntimeError: Cannot find CUDA_HOME. CUDA must be available to build the package.
#179
opened May 25, 2025 by
wujpia
wonder why using different(128/64) block-size in sageattn-v1 for quant q, k
#167
opened May 8, 2025 by
ZJLi2013
How to simply verify the successful installation of sageattention?
#165
opened May 6, 2025 by
aswordok
BUG: RTX 50XX nan returned by _fused.mean_scale_fuse_quant_cuda and _fused.scale_fuse_quant_cuda
#164
opened Apr 30, 2025 by
deepbeepmeep
The accuracy loss in the CUDA version is much more than Triton version for llama-3.2
#154
opened Apr 7, 2025 by
WanliZhong
K Sampler [WinError 2] The system cannot find the file specified.
#152
opened Apr 2, 2025 by
REG-0422
SageAttention fails to work on RTX 50xx series GPU (e.g. 5090) despite clean venv install
#148
opened Mar 24, 2025 by
Teskun
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.