ruvector_attention/sparse/
mod.rs

1//! Sparse attention mechanisms for efficient computation on long sequences
2//!
3//! This module provides sparse attention patterns that reduce complexity from O(n²) to sub-quadratic.
4
5pub mod flash;
6pub mod linear;
7pub mod local_global;
8pub mod mask;
9
10pub use flash::FlashAttention;
11pub use linear::LinearAttention;
12pub use local_global::LocalGlobalAttention;
13pub use mask::{AttentionMask, SparseMaskBuilder};