Skip to content

Implements dual-chunk-flash-attn backend for dual chunk attention with sparse attention support#11844

Merged
simon-mo merged 1 commit intovllm-project:mainfrom
sighingnow:dev/dual-chunk-attn
May 13, 2025

Commits

Commits on May 10, 2025