-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Issues: Dao-AILab/flash-attention
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
ERROR: No matching distribution found for flash-attn==2.6.3+cu123torch2.4cxx11abifalse
#1423
opened Jan 6, 2025 by
carolynsoo
Unable to install
flash-attn
even if I first install torch
alone
#1421
opened Jan 3, 2025 by
ytxmobile98
Is there a plan to support flash_attn_varlen_backward with fp8
#1420
opened Jan 3, 2025 by
gaodaheng
flash_attn_with_kvcache
discrepancy slicing kv_cache / cache_seqlens
#1417
opened Jan 1, 2025 by
jeromeku
looking for a test to compare the result with the KV cache updated in place and without the KV cache
#1414
opened Dec 26, 2024 by
chakpongchung
Performance Impact of Using Three Warps per Group (WG) in FA3 Compared to Two WGs
#1413
opened Dec 24, 2024 by
ziyuhuang123
UnboundLocalError: local variable 'out' referenced before assignment
#1412
opened Dec 24, 2024 by
chuangzhidan
Why Does FA3 Use Registers Instead of Directly Accessing SMEM with WGMMA on SM90?
#1407
opened Dec 23, 2024 by
ziyuhuang123
is flash_attn_with_kvcache() supposed to work for seqlen > 1 ?
#1402
opened Dec 20, 2024 by
vince62s
Understanding the Role of arrive in NamedBarrier Synchronization
#1400
opened Dec 19, 2024 by
ziyuhuang123
Why Doesn't FlashAttention3 Allow KV and O to Share Memory Space?
#1396
opened Dec 18, 2024 by
ziyuhuang123
g2s K tensor when handling padding in the seq_k, clear it rather than keeping the default SMEM values.
#1395
opened Dec 18, 2024 by
NVIDIA-JerryChen
Why does NamedBarrier in epilogue use NumMmaThreads(256) + NumThreadsPerWarp(32)?
#1389
opened Dec 16, 2024 by
ziyuhuang123
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.