-
Notifications
You must be signed in to change notification settings - Fork 44
Issues: pytorch-labs/attention-gym
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
RuntimeError in flex_attention: CUDA Device-Side Assertion Failure
#130
opened Mar 21, 2025 by
NuanBaobao
Would FlexAttention be useful over SDPA for masked language modeling?
#129
opened Mar 20, 2025 by
abdulfatir
Can FlexAttention Optimize Masks for Large Table Constraints?
#106
opened Jan 15, 2025 by
RaphaelMouravieff
FlexAttention uses much more GPU memory than FlashAttention-2
#101
opened Jan 9, 2025 by
ChenlongDeng
Illegal memory access on backward when there are unused block masks (nightly build)
#96
opened Dec 28, 2024 by
timt51
Short vs long sequences performance
question
Further information is requested
#89
opened Dec 12, 2024 by
francoishernandez
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.