use std::marker::PhantomData;
use crate::components::global::{
multi_stage::LoadMaxRoundPlaneCount, read::async_copy::async_copy_from,
};
use crate::{
components::global::read::{
FullLoadingStrategy, async_barrier::AsyncCopy, async_copy::ASYNC_COPY_WIDTH,
tiled::TiledLayout,
},
launch::RuntimeConfig,
};
use crate::{
components::global::read::{validate_async_barrier, validate_swizzle_atom_size},
components::global::read::{validate_async_copy, validate_async_copy_with_problem},
components::global::{GlobalReaderConfig, PlaneFlowPartition},
};
use crate::{
components::stage::StridedStageFamily,
components::stage::{ContiguousTilingLayout, StridedStageMemory, TilingOrder},
components::{global::memory::GlobalIterator, stage::TilingValidation},
definition::{MatmulElems, MatmulProblem, StageIdent},
};
use cubecl::{
prelude::*,
std::tensor::layout::{Layout, LayoutExpand},
{ir::DeviceProperties, prelude::barrier::Barrier},
};
use cubek_std::{InvalidConfigError, tile::Strided};
use super::{LoadingJob, LoadingValidation, ReaderMode};
#[derive(CubeType, Clone, Copy)]
pub struct AsyncFullCyclicLoading<T: TilingOrder> {
#[cube(comptime)]
_t: PhantomData<T>,
}
impl<TO: TilingOrder> LoadingValidation for AsyncFullCyclicLoading<TO> {
fn validate_with_config(
device_props: &DeviceProperties,
config: &GlobalReaderConfig,
) -> Result<(), InvalidConfigError> {
let vector_size = ASYNC_COPY_WIDTH / config.smem_config.dtype.size_bits() as u32;
if let ReaderMode::Strict = config.reader_mode {
let num_stage_vectors = config.smem_config.elements_per_stage() / vector_size;
let total_units = config.loading_units_count();
if !num_stage_vectors.is_multiple_of(total_units) {
return Err(Box::new(format!(
"Too many data will be loaded, resulting in out of bounds.
Try setting vector size and number of planes so that total unit count {total_units:?} divides number of vectors in stage.",
)));
}
}
if !config
.smem_config
.elements_per_tile_along_contiguous_dim()
.is_multiple_of(vector_size)
{
return Err(Box::new("Tile size isn't divisible by copy vector size"));
}
validate_swizzle_atom_size(config.smem_config)?;
validate_async_barrier(device_props)?;
validate_async_copy(
device_props,
&config.gmem_config.dtype,
&config.smem_config.dtype,
)?;
ContiguousTilingLayout::<TO>::check(config.smem_config)?;
Ok(())
}
fn validate_with_problem(
problem: &MatmulProblem,
dtypes: &MatmulElems,
ident: StageIdent,
) -> Result<(), InvalidConfigError> {
validate_async_copy_with_problem(problem, dtypes, ident)
}
}
impl<TO: TilingOrder> LoadMaxRoundPlaneCount for AsyncFullCyclicLoading<TO> {
fn max_round_plane_count(
elements_per_tile: u32,
tiles_per_stage: u32,
_vector_size: VectorSize,
plane_dim: u32,
dtype: StorageType,
) -> u32 {
let vector_size = ASYNC_COPY_WIDTH / dtype.size_bits() as u32;
let elements_per_stage = elements_per_tile * tiles_per_stage;
let num_vectors = elements_per_stage / vector_size;
num_vectors.div_ceil(plane_dim)
}
}
#[cube]
impl<TO: TilingOrder, RC: RuntimeConfig> FullLoadingStrategy<RC> for AsyncFullCyclicLoading<TO> {
type TilingLayout = ContiguousTilingLayout<TO>;
type SyncStrategy = AsyncCopy;
type Job<EG: Numeric, NG: Size, ES: Numeric, NS: Size> = AsyncFullCyclicJob;
type Stage = StridedStageFamily;
type TileKind = Strided;
fn new_job<EG: Numeric, NG: Size, ES: Numeric, NS: Size>(
_runtime_config: RC,
#[comptime] config: GlobalReaderConfig,
) -> Self::Job<EG, NG, ES, NS> {
let type_size = ES::type_size_bits().comptime();
let vector_size = ASYNC_COPY_WIDTH / type_size as u32;
let tile_num_elements = config.smem_config.elements_per_tile();
let num_stage_elements = config.smem_config.elements_per_stage();
let num_stage_vectors = num_stage_elements.div_ceil(vector_size);
let total_units = config.loading_units_count();
let num_tasks_per_unit = num_stage_vectors.div_ceil(total_units);
let balanced_workload = num_stage_vectors.is_multiple_of(total_units);
let jump_length = total_units * vector_size;
let unit_id = PlaneFlowPartition::new(config.plane_flow_config.partition_rule)
.load_index(config.input_load_flow)
* config.plane_dim
+ UNIT_POS_X;
let unit_position_base = unit_id * vector_size;
AsyncFullCyclicJob {
unit_position_base,
num_tasks_per_unit,
tile_num_elements,
jump_length,
copy_vector_size: vector_size,
balanced_workload,
num_stage_elements,
reader_mode: config.reader_mode,
}
}
}
#[derive(CubeType, Clone, Copy)]
pub struct AsyncFullCyclicJob {
unit_position_base: u32,
#[cube(comptime)]
num_tasks_per_unit: u32,
#[cube(comptime)]
tile_num_elements: u32,
#[cube(comptime)]
jump_length: u32,
#[cube(comptime)]
copy_vector_size: u32,
#[cube(comptime)]
balanced_workload: bool,
#[cube(comptime)]
num_stage_elements: u32,
#[cube(comptime)]
reader_mode: ReaderMode,
}
#[cube]
impl<EG: Numeric, NG: Size, ES: Numeric, NS: Size, TO: TilingOrder>
LoadingJob<EG, NG, ES, NS, ContiguousTilingLayout<TO>, AsyncCopy> for AsyncFullCyclicJob
{
type Stage = StridedStageFamily;
fn execute_task(
this: &mut Self,
#[comptime] task_id: u32,
global_iter: &GlobalIterator<Vector<EG, NG>>,
stage: &mut StridedStageMemory<ES, NS, ContiguousTilingLayout<TO>>,
_barrier: &mut Shared<Barrier>,
#[comptime] config: GlobalReaderConfig,
) {
let unit_position = this.unit_position_base + task_id * this.jump_length;
#[allow(clippy::collapsible_else_if)]
if comptime!(this.reader_mode == ReaderMode::Strict || this.balanced_workload) {
copy_vector::<EG, NG, ES, NS, TO>(this, unit_position, global_iter, stage, config);
} else {
if unit_position < this.num_stage_elements {
copy_vector::<EG, NG, ES, NS, TO>(this, unit_position, global_iter, stage, config);
}
}
}
fn task_count(this: &Self) -> comptime_type!(u32) {
this.num_tasks_per_unit
}
}
#[cube]
pub(crate) fn copy_vector<EG: Numeric, NG: Size, ES: Numeric, NS: Size, TO: TilingOrder>(
job: &AsyncFullCyclicJob,
unit_position: u32,
global_iter: &GlobalIterator<Vector<EG, NG>>,
stage: &mut StridedStageMemory<ES, NS, ContiguousTilingLayout<TO>>,
#[comptime] config: GlobalReaderConfig,
) {
let nth_tile = unit_position / job.tile_num_elements;
let pos_within_tile = unit_position % job.tile_num_elements;
let layout = TiledLayout::new(config.stage_ident, config.smem_config);
let view = global_iter.view();
let tile = ContiguousTilingLayout::<TO>::to_x_y(nth_tile, config.smem_config);
let pos = layout.to_source_pos((tile, pos_within_tile));
let stage_offset = unit_position / stage.smem.vector_size() as u32;
async_copy_from(view, pos, stage, stage_offset, config, job.copy_vector_size);
}