-
Notifications
You must be signed in to change notification settings - Fork 34
Issues: pytorch-labs/attention-gym
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Feature Request: Support for Dynamic Bias Tensor in FlexAttention Without Recompilation
#123
opened Feb 18, 2025 by
pengzhangzhi
Can FlexAttention Optimize Masks for Large Table Constraints?
#106
opened Jan 15, 2025 by
RaphaelMouravieff
FlexAttention uses much more GPU memory than FlashAttention-2
#101
opened Jan 9, 2025 by
ChenlongDeng
Illegal memory access on backward when there are unused block masks (nightly build)
#96
opened Dec 28, 2024 by
timt51
Short vs long sequences performance
question
Further information is requested
#89
opened Dec 12, 2024 by
francoishernandez
How to implement Bidirectional Alibi with padding using flex attention?
#74
opened Nov 7, 2024 by
sphmel
Is there any chance to call backward function dircetly instead of using pytorch autograd mechanism?
#73
opened Nov 7, 2024 by
MayDomine
Previous Next
ProTip!
Follow long discussions with comments:>50.