fla-org/flash-linear-attention
Efficient Triton-based linear attention kernels for PyTorch and multiple hardware platforms

View on index · View in 3D Map
// SURVEILLANCE FEED
Discovered repositories from the open source frontier
Efficient Triton-based linear attention kernels for PyTorch and multiple hardware platforms

View on index · View in 3D Map