1use tract_linalg::mmm::{EagerPackedInput, MMMInputValue, MatMatMul, PackedOpaqueFact};
2use tract_linalg::pack::{PackedFormat, PackingWriter};
3
4use crate::internal::*;
5use ndarray::prelude::*;
6use num_integer::Integer;
7
8use crate::ops::cnn::pools::{ConcretePoolGeometry, PoolGeometry};
9use crate::ops::cnn::{GeometryBound, PoolSpec, ResolveTo};
10use crate::ops::nn::{BaseDataShape, DataFormat, DataShape};
11
12#[derive(Debug, Clone, PartialEq, Eq, Hash)]
13pub struct Im2Col {
14 pub pool_spec: PoolSpec,
15 pub group: usize,
16 geometry: GeometryBound<SymbolicGeometry, ConcreteGeometry>,
17}
18
19#[derive(Debug, Clone, Hash, PartialEq, Eq)]
20struct SymbolicGeometry {
21 group: usize,
22 pool_spec: PoolSpec,
23 pool_geometry: PoolGeometry,
24 b_pack: PackedFormat,
25 k: usize,
26}
27
28#[derive(Debug, Clone, Hash, PartialEq, Eq)]
29struct ConcreteGeometry {
30 pool: ConcretePoolGeometry,
31 pub n: usize,
32 k: usize,
33 pub b_pack: PackedFormat,
34 pub ci_per_group: usize,
35 patcher: Patcher,
36 input_shape_with_n: DataShape,
37 packed_shape: TVec<usize>, }
39
40impl GeometryBound<SymbolicGeometry, ConcreteGeometry> {
41 pub fn b_pack(&self) -> &PackedFormat {
42 match self {
43 GeometryBound::Symbolic(s) => &s.b_pack,
44 GeometryBound::Concrete(s) => &s.b_pack,
45 }
46 }
47 pub fn k(&self) -> usize {
48 match self {
49 GeometryBound::Symbolic(s) => s.k,
50 GeometryBound::Concrete(s) => s.k,
51 }
52 }
53}
54
55impl ResolveTo<ConcreteGeometry> for SymbolicGeometry {
56 type Param = [usize];
57 fn resolve(&self, input_full_shape: &[usize]) -> TractResult<ConcreteGeometry> {
58 let pool = self.pool_geometry.to_concrete(input_full_shape)?.into_owned();
59 let patcher = if !pool.patch.padded && pool.patch.rank() == 2 {
60 Patcher::Valid2d
61 } else if pool.patch.rank() == 2 {
62 Patcher::Padded2d
63 } else if !pool.patch.padded && pool.patch.rank() == 1 {
64 Patcher::Valid1d
65 } else {
66 Patcher::Generic
67 };
68 let ci_per_group = pool.input_shape.c_dim() / self.group;
69 let n = pool.output_shape.hw_dims().iter().product();
70 let input_shape_with_n = match self.pool_spec.data_format {
71 DataFormat::HWC => DataFormat::NHWC.from_n_c_hw(
72 1,
73 *pool.input_shape.c(),
74 pool.input_shape.hw_dims(),
75 )?,
76 DataFormat::CHW => DataFormat::NCHW.from_n_c_hw(
77 1,
78 *pool.input_shape.c(),
79 pool.input_shape.hw_dims(),
80 )?,
81 _ => pool.input_shape.clone(),
82 };
83 let packed_shape = Im2Col::packed_shape(&pool.input_shape, self.group)?;
84 Ok(ConcreteGeometry {
85 pool,
86 n,
87 k: self.k,
88 ci_per_group,
89 b_pack: self.b_pack.clone(),
90 patcher,
91 input_shape_with_n,
92 packed_shape,
93 })
94 }
95}
96
97impl Im2Col {
98 pub fn new(
99 pool_spec: PoolSpec,
100 group: usize,
101 k: usize,
102 input_full_shape: &ShapeFact,
103 mmm: Box<dyn MatMatMul>,
104 packing: usize,
105 ) -> TractResult<Im2Col> {
106 let b_pack = mmm.packings()[packing]
107 .1
108 .downcast_ref::<PackedFormat>()
109 .context("Im2Col expects regular packed format")?
110 .clone();
111
112 let pool_geometry = pool_spec.compute_geo(input_full_shape)?;
113 let geometry: GeometryBound<_, _> =
114 SymbolicGeometry { group, pool_spec: pool_spec.clone(), pool_geometry, b_pack, k }
115 .into();
116 let geometry = geometry.optimize_if(input_full_shape.as_concrete())?;
117 Ok(Im2Col { pool_spec, group, geometry })
118 }
119
120 fn packed_shape<D: DimLike>(
122 input_shape: &BaseDataShape<D, TVec<D>>,
123 group: usize,
124 ) -> TractResult<TVec<D>> {
125 let mut output_shape: TVec<D> = tvec!();
126 output_shape.push(input_shape.n().cloned().unwrap_or_else(|| 1.into()));
127 output_shape.push(group.into());
128 Ok(output_shape)
129 }
130}
131
132impl Op for Im2Col {
133 fn name(&self) -> StaticName {
134 "Im2col".into()
135 }
136
137 fn info(&self) -> TractResult<Vec<String>> {
138 Ok(vec![format!("groups:{}", self.group)])
139 }
140
141 impl_op_same_as!();
142 op_as_typed_op!();
143}
144
145impl EvalOp for Im2Col {
146 fn is_stateless(&self) -> bool {
147 true
148 }
149
150 fn eval(&self, mut inputs: TVec<TValue>) -> TractResult<TVec<TValue>> {
151 let geometry = self.geometry.to_concrete(inputs[0].shape())?;
152 unsafe {
153 let mut input = inputs.remove(0).into_tensor();
154 let pad_value: Option<&Tensor> = if inputs.len() > 0 { Some(&inputs[0]) } else { None };
155 let mut output = Tensor::uninitialized::<Opaque>(&geometry.packed_shape)?;
156 if !self.pool_spec.data_format.has_n() {
157 input.insert_axis(0)?;
158 }
159 let mut output_view = output.to_array_view_mut::<Opaque>()?;
160 let panel_bytes =
161 geometry.b_pack.single_panel_len(geometry.k) * input.datum_type().size_of();
162
163 if !geometry.pool.output_shape.shape.contains(&0) {
166 for i in 0..*geometry.input_shape_with_n.n().unwrap_or(&1) {
167 let input = input.view_at_prefix(&[i])?;
168 for g in 0..self.group {
169 let mut data = Tensor::uninitialized_aligned_dt(
170 input.datum_type(),
171 &[geometry.b_pack.len(geometry.k, geometry.n)],
172 geometry.b_pack.alignment(),
173 )?;
174 dispatch_copy_by_size!(Patcher::patch(input.datum_type())(
175 &geometry.patcher,
176 &geometry,
177 &input,
178 &mut data.view_mut(),
179 g,
180 pad_value
181 ))?;
182 let input: Box<dyn MMMInputValue> = Box::new(EagerPackedInput {
183 fact: PackedOpaqueFact {
184 format: Box::new(geometry.b_pack.clone()),
185 k: geometry.k,
186 mn: geometry.n.to_dim(),
187 },
188 packed: data.into_blob()?.into(),
189 panel_bytes,
190 mn: geometry.n,
191 });
192 output_view[[i, g]] = input.into();
193 }
194 }
195 }
196 Ok(tvec!(output.into_tvalue()))
197 }
198 }
199}
200
201impl TypedOp for Im2Col {
202 as_op!();
203
204 fn output_facts(&self, inputs: &[&TypedFact]) -> TractResult<TVec<TypedFact>> {
205 let input_shape = self.pool_spec.data_format.shape(inputs[0].shape.to_tvec())?;
206 let output_shape = self.pool_spec.output_shape(&inputs[0].shape)?;
207 let mn = output_shape.hw_dims().iter().product::<TDim>();
208 let pof = PackedOpaqueFact {
209 format: Box::new(self.geometry.b_pack().clone()),
210 k: self.geometry.k(),
211 mn,
212 };
213 Ok(tvec!(
214 Opaque::fact(&[input_shape.n().cloned().unwrap_or(1.into()), self.group.into()])
215 .with_opaque_fact(pof)
216 ))
217 }
218
219 fn declutter(
220 &self,
221 model: &TypedModel,
222 node: &TypedNode,
223 ) -> TractResult<Option<TypedModelPatch>> {
224 let input_fact = model.outlet_fact(node.inputs[0])?;
225 if node.inputs.len() == 2
226 && model.outlet_fact(node.inputs[1])?.konst.as_ref().and_then(|t| t.as_uniform())
227 == Some(Tensor::zero_scalar_dt(input_fact.datum_type)?)
228 {
229 Ok(Some(
230 TypedModelPatch::replace_single_op(model, node, &node.inputs[0..1], self.clone())?
231 .with_context("b0 is zero"),
232 ))
233 } else {
234 Ok(None)
235 }
236 }
237}
238
239#[derive(Copy, Clone, Debug, Hash, PartialEq, Eq)]
240enum Patcher {
241 Generic,
242 Valid1d,
243 Valid2d,
244 Padded2d,
245}
246
247impl Patcher {
248 fn patch<'p, T: Copy + Datum + num_traits::Zero>(
249 &self,
250 geo: &'p ConcreteGeometry,
251 input: &TensorView,
252 pack: &'p mut TensorView,
253 g: usize,
254 pad_value: Option<&Tensor>,
255 ) -> TractResult<()> {
256 match self {
257 Patcher::Valid1d => Self::valid_1d::<T>(geo, input, pack, g),
258 Patcher::Valid2d => Self::valid_2d::<T>(geo, input, pack, g),
259 Patcher::Padded2d => Self::padded_2d::<T>(
260 geo,
261 input,
262 pack,
263 g,
264 pad_value.unwrap_or(&Tensor::zero_scalar::<T>()?),
265 ),
266 _ => Self::generic::<T>(
267 geo,
268 input,
269 pack,
270 g,
271 pad_value.unwrap_or(&Tensor::zero_scalar::<T>()?),
272 ),
273 }
274 }
275
276 #[inline(never)]
277 fn generic<'p, T: Copy + Datum>(
278 geometry: &'p ConcreteGeometry,
279 input: &TensorView,
280 pack: &'p mut TensorView,
281 g: usize,
282 pad_value: &Tensor,
283 ) -> TractResult<()> {
284 unsafe {
285 let pad_value = *pad_value.to_scalar_unchecked();
286 let mut mega_matrix = Tensor::uninitialized::<T>(&[geometry.k, geometry.n])?;
287 let mut mega_matrix_view = mega_matrix.to_array_view_mut_unchecked::<T>();
288 let ptr = input.as_ptr_unchecked::<T>();
289 let ptr = ptr.add(geometry.input_shape_with_n.c_stride() * (g * geometry.ci_per_group));
290 for (spatial, mut col) in ndarray::indices(&*geometry.pool.patch.output_shape)
291 .into_iter()
292 .zip(mega_matrix_view.axis_iter_mut(Axis(1)))
293 {
294 let mut col = col.iter_mut();
295 for ci in 0..geometry.ci_per_group {
296 let ptr = ptr.add(geometry.input_shape_with_n.c_stride() * ci);
297 for v in geometry.pool.patch.at(spatial.slice()) {
298 *col.next().expect("geometry error in conv") =
299 v.map(|o| *ptr.offset(o)).unwrap_or(pad_value);
300 }
301 }
302 }
303 geometry.b_pack.pack(pack, mega_matrix.view(), 0, 1);
304 Ok(())
305 }
306 }
307
308 #[inline(never)]
309 fn valid_1d<'p, T: Copy + Datum>(
310 geometry: &'p ConcreteGeometry,
311 input: &TensorView,
312 pack: &'p mut TensorView,
313 g: usize,
314 ) -> TractResult<()> {
315 unsafe {
316 let x_stride = *geometry.input_shape_with_n.h_stride() as isize
317 * geometry.pool.patch.spec.strides[0] as isize;
318 let c_stride = *geometry.input_shape_with_n.c_stride() as isize;
319 let pack = pack.as_slice_mut_unchecked::<T>();
320 let mut writer =
321 geometry.b_pack.write_with_k_outer(pack.as_mut_ptr(), geometry.k, geometry.n);
322 let iptr = input.as_ptr_unchecked::<T>();
323 let iptr = iptr.add(g * geometry.ci_per_group * geometry.input_shape_with_n.c_stride());
324 for ci in 0..geometry.ci_per_group {
325 let iptr = iptr.offset(ci as isize * c_stride);
326 for koffset in &geometry.pool.patch.standard_layout_data_field {
327 let iptr = iptr.offset(*koffset);
328 for x in 0..*geometry.pool.patch.output_shape.get_unchecked(0) {
329 writer.write(*iptr.offset(x as isize * x_stride));
330 }
331 }
332 }
333 Ok(())
334 }
335 }
336
337 #[inline(never)]
338 fn padded_2d<'p, T: Copy + Datum>(
339 geometry: &'p ConcreteGeometry,
340 input: &TensorView,
341 pack: &'p mut TensorView,
342 g: usize,
343 pad_value: &Tensor,
344 ) -> TractResult<()> {
345 unsafe {
346 let pad_value = *pad_value.to_scalar_unchecked();
347 let pack = pack.as_slice_mut_unchecked::<T>();
348 let y_stride = geometry.pool.patch.spec.strides[0] as isize;
349 let x_stride = geometry.pool.patch.spec.strides[1] as isize;
350 let shape = &geometry.input_shape_with_n;
351 let y_stride_ptr = y_stride * *shape.h_stride() as isize;
352 let x_stride_ptr = x_stride * *shape.w_stride() as isize;
353 let c_stride_ptr = *shape.c_stride() as isize;
354 let input_heigth = shape.hw_dims()[0] as isize;
355 let input_width = shape.hw_dims()[1] as isize;
356 let kernel_len = geometry.pool.patch.standard_layout_data_field.len();
357 let mut writer =
358 geometry.b_pack.write_with_k_outer(pack.as_mut_ptr(), geometry.k, geometry.n);
359 let iptr = input.as_ptr_unchecked::<T>();
360 let iptr = iptr.add(g * geometry.ci_per_group * shape.c_stride());
361 let output_width = *geometry.pool.patch.output_shape.get_unchecked(1);
362 for ci in 0..geometry.ci_per_group {
363 let iptr = iptr.offset(ci as isize * c_stride_ptr);
364 for kitem in 0..kernel_len {
365 let dy = *geometry.pool.patch.data_field.as_ptr().offset(kitem as isize * 2);
366 let dx =
367 *geometry.pool.patch.data_field.as_ptr().offset(1 + kitem as isize * 2);
368 let valid_x_start =
369 Integer::div_ceil(&-dx, &x_stride).max(0).min(output_width as _);
370 let valid_x_end = Integer::div_ceil(&(input_width - dx), &x_stride)
371 .max(0)
372 .min(output_width as _);
373
374 let iptr = iptr.offset(
375 *geometry.pool.patch.standard_layout_data_field.get_unchecked(kitem),
376 );
377 for yo in 0..*geometry.pool.patch.output_shape.get_unchecked(0) {
378 let y = yo as isize * y_stride + dy;
379 let iptr = iptr.offset(yo as isize * y_stride_ptr);
380 if y >= 0 && y < input_heigth {
381 Self::padded_2d_invalid_x_loop(
382 valid_x_start as usize,
383 pad_value,
384 &mut writer,
385 );
386 Self::padded_2d_valid_x_loop(
387 valid_x_start,
388 valid_x_end,
389 x_stride_ptr,
390 iptr,
391 &mut writer,
392 );
393 Self::padded_2d_invalid_x_loop(
394 output_width - valid_x_end as usize,
395 pad_value,
396 &mut writer,
397 );
398 } else {
399 Self::padded_2d_invalid_x_loop(output_width, pad_value, &mut writer);
400 }
401 }
402 }
403 }
404 }
405 Ok(())
406 }
407
408 #[inline(never)]
409 unsafe fn padded_2d_invalid_x_loop<T: Copy + Datum>(
410 count: usize,
411 pad_value: T,
412 writer: &mut tract_linalg::pack::KOutWriter<T>,
413 ) {
414 for _ in 0..count {
415 writer.write(pad_value);
416 }
417 }
418
419 #[inline(never)]
420 unsafe fn padded_2d_valid_x_loop<T: Copy + Datum>(
421 x_min: isize,
422 x_max: isize,
423 x_stride_ptr: isize,
424 iptr: *const T,
425 writer: &mut tract_linalg::pack::KOutWriter<T>,
426 ) {
427 for x in x_min..x_max {
428 writer.write(unsafe { *iptr.offset(x * x_stride_ptr) });
429 }
430 }
431
432 #[inline(never)]
433 fn valid_2d<'p, T: Copy + Datum>(
434 geometry: &'p ConcreteGeometry,
435 input: &TensorView,
436 pack: &'p mut TensorView,
437 g: usize,
438 ) -> TractResult<()> {
439 unsafe {
440 let pack = pack.as_slice_mut_unchecked::<T>();
441 let shape = &geometry.input_shape_with_n;
442 let y_stride = geometry.pool.patch.spec.strides[0] as isize;
443 let x_stride = geometry.pool.patch.spec.strides[1] as isize;
444 let y_stride_ptr = y_stride * *shape.h_stride() as isize;
445 let x_stride_ptr = x_stride * *shape.w_stride() as isize;
446 let c_stride_ptr = *shape.c_stride() as isize;
447 let mut writer =
448 geometry.b_pack.write_with_k_outer(pack.as_mut_ptr(), geometry.k, geometry.n);
449 let iptr = input.as_ptr_unchecked::<T>();
450 let iptr = iptr.add(g * geometry.ci_per_group * shape.c_stride());
451 for ci in 0..geometry.ci_per_group {
452 let iptr = iptr.offset(ci as isize * c_stride_ptr);
453 for koffset in &geometry.pool.patch.standard_layout_data_field {
454 let iptr = iptr.offset(*koffset);
455 for y in 0..*geometry.pool.patch.output_shape.get_unchecked(0) {
456 let iptr = iptr.offset(y as isize * y_stride_ptr);
457 for x in 0..*geometry.pool.patch.output_shape.get_unchecked(1) {
458 writer.write(*iptr.offset(x as isize * x_stride_ptr));
459 }
460 }
461 }
462 }
463 Ok(())
464 }
465 }
466}