-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Issues: triton-lang/triton
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
runtime JIT execution of the same kernel incurs high overhead(binder function)
performance
#6064
opened Feb 28, 2025 by
Learnmore666
Triton incorrectly interprets uint8 indices as int8 in pointer arithmetic operations
bug
#6043
opened Feb 27, 2025 by
Ivan1248
fused attention (flash attention v2) doesn't support bfloat16?
#6005
opened Feb 24, 2025 by
guanzhchen
tl.sort and torch.sort give inconsistent results when input contains inf or nan
bug
#5999
opened Feb 24, 2025 by
chenmiao1919
Bug in tutorials/06-fused-attention.py: test_op assertion fails for specific input.
bug
#5971
opened Feb 20, 2025 by
p81sunshine
Dose Triton supports new features of Blackwell for RTX5090 and 5080?
#5950
opened Feb 18, 2025 by
jt-zhang
Triton kernel not compiling with multiple threads and GPUs
bug
#5933
opened Feb 15, 2025 by
amnamasood-amd
Upstream LLVM SLP vectorizer change requires the correct triple
performance
#5930
opened Feb 14, 2025 by
pclove1
f16 -> f8e5m2 conversion on H100 does not preserve infinities
bug
#5927
opened Feb 14, 2025 by
bchetioui
RTNE semantics are not respected by
tt.fp_to_fp
going from f16 to f8e5m2 on A100
bug
#5925
opened Feb 14, 2025 by
bchetioui
SystemError: PY_SSIZE_T_CLEAN macro must be defined for '#' formats
bug
#5919
opened Feb 14, 2025 by
famiu
Performance Discrepancy Between Matrix Multiplication written in 3D and 2D Loads
performance
#5916
opened Feb 13, 2025 by
nullplay
Previous Next
ProTip!
Follow long discussions with comments:>50.