Skip to content

Conversation

xiangze-arm
Copy link

  • Implement flash attention for context attention
  • Implement flash decoding for decoder self attention
  • Avoid cache assem and use blocked kv cache directly
  • Compute GQA by group of heads in flash decoding

- Implement flash attention for context attention
- Implement flash decoding for decoder self attention
- Avoid cache assem and use blocked kv cache directly
- Compute GQA by group of heads in flash decoding

Signed-off-by: Zhang Xiangze <[email protected]>
Co-authored-by: Ruifeng Wang <[email protected]>
@CLAassistant
Copy link

CLAassistant commented Aug 20, 2025

CLA assistant check
All committers have signed the CLA.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants