Change the repository type filter
All
Repositories list
6 repositories
flash-linear-attention
Public🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton- Flash-Linear-Attention models beyond language
- 🔥 A minimal training framework for scaling FLA models
flash-hybrid-attention
Public- Triton implement of bi-directional (non-causal) linear attention