//! Attention-specific interpretation of the generic [CubeMapping] from `cubek-std`.
//!
//! Attention has 2D problem-space axes: `(seq_q_tile, batch_heads)` where
//! `batch_heads = batch * num_heads`. The [HypercubeBlueprint], [CubeCountPlan]
//! and [CubeMapping] types come directly from `cubek-std`; this module only
//! adds the attention-specific `(seq_q, batch_heads)` mapper.
use cubecl;
use *;
pub use ;
/// Reads the cube position as attention `(seq_q_index, batch_heads_index)` coordinates.
///
/// The `batch_heads_index` spans `batch * num_heads`; the third axis is unused.