Skip to main content

tract_core/ops/cnn/conv/
conv.rs

1use tract_data::itertools::izip;
2use tract_linalg::WeightType;
3use tract_linalg::block_quant::{BlockQuantFact, PackedBlockQuantFormat};
4use tract_num_traits::Zero;
5
6use crate::internal::*;
7use crate::model::*;
8use crate::ops;
9use crate::ops::array::Pad;
10use crate::ops::array::PadMode;
11use crate::ops::binary::TypedBinOp;
12use crate::ops::cast::cast;
13use crate::ops::cnn::PaddingSpec::*;
14use crate::ops::cnn::conv::block_quant::{BlockQuantIntoShape, SplitGroupBlockQuant};
15use crate::ops::cnn::conv::lazy_im2col::LazyIm2Col;
16use crate::ops::cnn::conv::lazy_im2col::LazyIm2colParams;
17use crate::ops::cnn::wire_reshape_bias_for_bin;
18use crate::ops::einsum::EinSum;
19use crate::ops::math::{Add, Div, Mul, Sub};
20use crate::ops::math::{add, div, mul, sub};
21use crate::ops::matmul::ModePicker;
22use crate::ops::matmul::optimized::AddMatMulGeometry;
23use crate::ops::matmul::optimized::MapOutputAxisToInput;
24use crate::ops::matmul::pack::{OptMatMulPack, OptSimpleMatMulPack};
25use crate::ops::matmul::quant::wire_ensure_q8_flavour;
26use crate::ops::nn::Reduce;
27
28use super::depth_wise::DepthWise;
29use super::im2col::Im2Col;
30use crate::ops::cnn::conv::{KernelFormat, block_quant_aware_weight_shape};
31use crate::ops::cnn::pools::{ConcretePoolGeometry, PoolGeometry, PoolSpec};
32use crate::ops::matmul::optimized::{OptMatMul, ProtoFusedSpec};
33use crate::ops::nn::{BaseDataShape, DataFormat, DataShape};
34
35use tract_linalg::mmm::{MMMInputFormat, MatMatMul};
36use tract_linalg::pack::PackedFormat;
37
38#[derive(Debug, Clone, new, Hash)]
39pub struct Conv {
40    pub pool_spec: PoolSpec,
41    pub kernel_fmt: KernelFormat,
42    pub group: usize,
43    // None -> floats
44    // Some(I32) -> output is I32 (use quantized kernels, but output will be i32). last 2 Q inputs
45    // are ignored
46    // Some(QXX) -> quantized XX, but parameters are ignored (I8, U8, or I32) in favor of last 2 Q inputs
47    pub q_params: Option<DatumType>,
48}
49
50impl Conv {
51    pub fn input_channels(&self) -> usize {
52        self.pool_spec.input_channels
53    }
54
55    pub fn output_channels(&self) -> usize {
56        self.pool_spec.output_channels
57    }
58
59    pub fn wire_kernel_as_g_o_ihw(
60        &self,
61        model: &mut TypedModel,
62        name: &str,
63        mut kernel: OutletId,
64    ) -> TractResult<TVec<OutletId>> {
65        let fact = model.outlet_fact(kernel)?;
66        if fact.datum_type.is_opaque() {
67            ensure!(self.kernel_fmt == KernelFormat::OIHW && fact.rank() == 0);
68            kernel = model.wire_node(
69                format!("{name}.prep_kernel.g"),
70                SplitGroupBlockQuant { group: self.group },
71                &[kernel],
72            )?[0];
73            kernel = model.wire_node(
74                format!("{name}.prep_kernel.ihw"),
75                BlockQuantIntoShape {
76                    shape: tvec!(
77                        self.output_channels() / self.group,
78                        self.input_channels() / self.group
79                            * self.pool_spec.kernel_shape.iter().product::<usize>(),
80                    ),
81                },
82                &[kernel],
83            )?[0];
84            Ok(tvec!(kernel))
85        } else {
86            for (ix, op) in self
87                .kernel_fmt
88                .kernel_as_group_o_ihw_ops(&fact.shape, self.group)
89                .into_iter()
90                .enumerate()
91            {
92                kernel = model.wire_node(format!("{name}.prep_kernel.{ix}"), op, &[kernel])?[0];
93            }
94            Ok(tvec!(kernel))
95        }
96    }
97
98    fn wire_pack_g_o_ihw(
99        &self,
100        model: &mut TypedModel,
101        name: &str,
102        format: &dyn MMMInputFormat,
103        kernel: OutletId,
104    ) -> TractResult<OutletId> {
105        let fact = model.outlet_fact(kernel)?;
106        let wire = if fact.datum_type.is_opaque() {
107            let fact = model
108                .outlet_fact(kernel)?
109                .opaque_fact
110                .as_ref()
111                .and_then(|of| of.downcast_ref::<BlockQuantFact>())
112                .context("Only manage BlockQuant")?;
113            model.wire_node(
114                format!("{name}.prep_kernel.pack"),
115                OptSimpleMatMulPack {
116                    packed_format: format
117                        .downcast_ref::<PackedBlockQuantFormat>()
118                        .context("Expect a block quant format")?
119                        .clone(),
120                    k: fact.k(),
121                    m: fact.m(),
122                },
123                &[kernel],
124            )?
125        } else {
126            let format = format
127                .downcast_ref::<PackedFormat>()
128                .context("Expect regular packing for numeric weights")?;
129            model.wire_node(
130                format!("{name}.prep_kernel.pack"),
131                OptMatMulPack {
132                    packers: vec![format.clone()],
133                    k_axis: 2,
134                    mn_axis: 1,
135                    mode_picker: ModePicker::Single,
136                },
137                &[kernel],
138            )?
139        };
140        Ok(wire[0])
141    }
142
143    // group,bias
144    fn wire_bias_as_non_linear(
145        &self,
146        model: &mut TypedModel,
147        name: &str,
148        bias: OutletId,
149        c_group_axis: usize,
150    ) -> TractResult<(ProtoFusedSpec, OutletId)> {
151        use tract_linalg::BinOp::Add;
152        let fact = model.outlet_fact(bias)?;
153        if fact.shape.volume().is_one() {
154            Ok((ProtoFusedSpec::BinScalar(2, Add), bias))
155        } else {
156            let bias = AxisOp::wire_split_axis(
157                model,
158                format!("{name}.reformat_bias"),
159                bias,
160                0,
161                self.group,
162            )?[0];
163            let pfs =
164                ProtoFusedSpec::BinPerRow(2, Add, MapOutputAxisToInput(tvec!((c_group_axis, 0))));
165            Ok((pfs, bias))
166        }
167    }
168
169    pub unsafe fn wire_as_quant_im2col(
170        &self,
171        model: &mut TypedModel,
172        name: &str,
173        wires: &[OutletId],
174    ) -> TractResult<TVec<OutletId>> {
175        ensure!(self.q_params.is_some());
176        use crate::ops::matmul::quant as qmm;
177
178        let c_dt = self.q_params.unwrap();
179        let &[mut x, mut kernel, bias, mut x0, x_scale, mut k0, mut k_scale, y0, y_scale] = wires
180        else {
181            bail!("Wrong number of inputs")
182        };
183        wire_ensure_q8_flavour(model, name, &mut kernel, "k", &mut k0, i8::datum_type())?;
184        wire_ensure_q8_flavour(model, name, &mut x, "x", &mut x0, i8::datum_type())?;
185
186        let a_fact = model.outlet_fact(kernel)?.clone();
187        let b_fact = model.outlet_fact(x)?.clone();
188
189        let (_geo, m, k, n) = self.compute_geo(&b_fact)?;
190        let (mmm, packing) = self.choose_impl(&b_fact, &a_fact, m, k, &n)?;
191        let output_shape = self.pool_spec.output_shape(&b_fact.shape)?;
192
193        if !model.outlet_fact(k_scale)?.shape.volume().is_one() {
194            // requant is performed before geo_reshape, so we need at most one geo axis to the
195            // right
196            if !output_shape.fmt.c_is_last() {
197                k_scale = model.wire_node(
198                    format!("{name}.a_scale_axis_fix"),
199                    AxisOp::Add(1),
200                    &[k_scale],
201                )?[0];
202            }
203        }
204
205        let abc_scale = qmm::combine_scales(model, name, k_scale, x_scale, y_scale)?;
206
207        let im2col = model.wire_node(
208            format!("{name}.im2col"),
209            Im2Col::new(
210                self.pool_spec.clone(),
211                self.group,
212                k,
213                &b_fact.shape,
214                mmm.clone(),
215                packing,
216            )?,
217            &[x, x0],
218        )?[0];
219
220        let g_o_ihw = self.wire_kernel_as_g_o_ihw(model, name, kernel)?;
221        let g_o_ihw_as_i32 =
222            model.wire_node(format!("{name}.kernel_as_i32"), cast(i32::datum_type()), &g_o_ihw)?;
223        let sum_ker_g_c_k = model.wire_node(
224            format!("{name}.sum_ker_g_c_k"),
225            Reduce::new(tvec!(2), ops::nn::Reducer::Sum),
226            &g_o_ihw_as_i32,
227        )?;
228        let sum_ker_a_g_c =
229            model.wire_node(format!("{name}.rm_k"), AxisOp::Rm(2), &sum_ker_g_c_k)?;
230        // align sum_A from G,C to "C" shape: N,HW,G,C (or N,G,C,HW)
231        let sum_ker_n_g_c = model.wire_node(
232            format!("{name}.sum_ker_n_g_c.axis_0"),
233            AxisOp::Add(0),
234            &sum_ker_a_g_c,
235        )?;
236        let hw_position = if self.pool_spec.data_format.c_is_last() { 1 } else { 3 };
237        let sum_ker = model.wire_node(
238            format!("{name}.sum_ker_n_g_c"),
239            AxisOp::Add(hw_position),
240            &sum_ker_n_g_c,
241        )?;
242
243        ensure!(mmm.packings()[packing].1.downcast_ref::<PackedFormat>().is_some());
244        let mut sum_x = model.wire_node(
245            format!("{name}.sum_x"),
246            super::QSumB { dt: b_fact.datum_type, n, r: mmm.nr(), k },
247            &[im2col],
248        )?;
249        // sum_b is N,G,HW. make it N,HW,G,C or N,G,C,HW
250        sum_x = model.wire_node(format!("{name}.add_c"), AxisOp::Add(2), &sum_x)?;
251        if self.pool_spec.data_format.c_is_last() {
252            sum_x =
253                model.wire_node(format!("{name}.transpose_sum_b"), AxisOp::Move(3, 1), &sum_x)?;
254        }
255
256        let (mmm_output_shape, c_axis, h_axis) = self.mmm_output_shape(&output_shape)?;
257        let bias_name = &model.node(bias.node).name;
258        let bias =
259            model.wire_node(format!("{bias_name}.cast"), cast(mmm.internal_type()), &[bias])?[0];
260        let wire = self.wire_mm_weights_bias(
261            model,
262            name,
263            im2col,
264            g_o_ihw[0],
265            bias,
266            mmm,
267            packing,
268            i32::datum_type(),
269            mmm_output_shape.clone().into(),
270            k,
271            c_axis,
272            h_axis,
273        )?;
274
275        let wire = qmm::compensate_zero_points(
276            model,
277            name,
278            wire[0],
279            k.to_dim(),
280            k0,
281            x0,
282            sum_ker[0],
283            sum_x[0],
284        )?;
285
286        let wire = self.wire_remove_group(model, name, &[wire], &mmm_output_shape, c_axis)?;
287        let wire = self.wire_rm_n_if_needed(model, name, &wire)?;
288        let wire = qmm::requant(model, name, wire[0], c_dt, abc_scale, y0)?;
289        Self::wire_geo_reshape(model, name, &[wire], &output_shape)
290    }
291
292    pub fn wire_remove_group<D: DimLike>(
293        &self,
294        model: &mut TypedModel,
295        name: &str,
296        wire: &[OutletId],
297        mmm_output_shape: &[D],
298        c_axis: usize,
299    ) -> TractResult<TVec<OutletId>> {
300        let m = &mmm_output_shape[c_axis];
301        let op = if self.group == 1 {
302            AxisOp::Rm(c_axis - 1)
303        } else {
304            AxisOp::Reshape(
305                c_axis - 1,
306                tvec!(self.group.to_dim(), m.to_dim()),
307                tvec!(m.to_dim() * self.group),
308            )
309        };
310        model.wire_node(format!("{name}.reshape_group"), op, wire)
311    }
312
313    pub unsafe fn wire_as_im2col_pair(
314        &self,
315        model: &mut TypedModel,
316        name: &str,
317        wire: &[OutletId],
318    ) -> TractResult<TVec<OutletId>> {
319        let &[x, w, bias] = wire else { bail!("Wrong number of inputs") };
320        let x_fact = model.outlet_fact(x)?.clone();
321        let w_fact = model.outlet_fact(w)?.clone();
322        let c_dt = crate::ops::matmul::output_type(x_fact.datum_type);
323
324        let (_, m, k, n) = self.compute_geo(&x_fact)?;
325        let (mmm, packing) = self.choose_impl(&x_fact, &w_fact, m, k, &n)?;
326        let geo_output_shape = self.pool_spec.output_shape(&x_fact.shape)?;
327        let (mmm_output_shape, c_axis, h_axis) = self.mmm_output_shape(&geo_output_shape)?;
328
329        let padding =
330            model.add_const(format!("{name}.b0"), Tensor::zero_scalar_dt(x_fact.datum_type)?)?;
331
332        let mut wire: TVec<_> = wire.into();
333        wire[0] = model.wire_node(
334            format!("{name}.im2col"),
335            Im2Col::new(
336                self.pool_spec.clone(),
337                self.group,
338                k,
339                &x_fact.shape,
340                mmm.clone(),
341                packing,
342            )?,
343            &[wire[0], padding],
344        )?[0];
345
346        let g_o_ihw = self.wire_kernel_as_g_o_ihw(model, name, wire[1])?;
347
348        let wire = self
349            .wire_mm_weights_bias(
350                model,
351                name,
352                wire[0],
353                g_o_ihw[0],
354                bias,
355                mmm,
356                packing,
357                c_dt,
358                mmm_output_shape.clone().into(),
359                k.to_usize().unwrap(),
360                c_axis,
361                h_axis,
362            )
363            .context("in wire_opt_matmul")?;
364
365        let wire = self.wire_remove_group(model, name, &wire, &mmm_output_shape, c_axis)?;
366        let wire = self.wire_rm_n_if_needed(model, name, &wire)?;
367        Self::wire_geo_reshape(model, name, &wire, &geo_output_shape)
368    }
369
370    // always have N and G. G is right before C, c_axis point to C, c_axis-1 points to G
371    fn mmm_output_shape<D: DimLike>(
372        &self,
373        output_shape: &BaseDataShape<D, TVec<D>>,
374    ) -> TractResult<(TVec<D>, usize, usize)> {
375        let geo_collapsed_out: D = output_shape.hw_dims().iter().cloned().product();
376        let shape: BaseDataShape<D, TVec<D>> = output_shape.fmt.with_n().from_n_c_hw(
377            output_shape.n().cloned().unwrap_or_else(|| 1.into()),
378            output_shape.c().clone(),
379            tvec!(geo_collapsed_out),
380        )?;
381        let mut mmm_output_shape: TVec<D> = shape.shape.clone();
382        let mut c_axis = shape.c_axis();
383        let mut h_axis = shape.h_axis();
384        mmm_output_shape[shape.c_axis()] = mmm_output_shape[c_axis].clone() / self.group;
385        mmm_output_shape.insert(c_axis, self.group.into());
386        if h_axis > c_axis {
387            h_axis += 1;
388        }
389        c_axis += 1;
390        Ok((mmm_output_shape, c_axis, h_axis))
391    }
392
393    fn wire_rm_n_if_needed(
394        &self,
395        model: &mut TypedModel,
396        name: &str,
397        wire: &[OutletId],
398    ) -> TractResult<TVec<OutletId>> {
399        if self.pool_spec.data_format.has_n() {
400            Ok(wire.into())
401        } else {
402            model.wire_node(format!("{name}.rm_n"), AxisOp::Rm(0), wire)
403        }
404    }
405
406    fn wire_geo_reshape<D: DimLike>(
407        model: &mut TypedModel,
408        name: &str,
409        wire: &[OutletId],
410        output_shape: &BaseDataShape<D, TVec<D>>,
411    ) -> TractResult<TVec<OutletId>> {
412        let geo_collapsed_out: D = output_shape.hw_dims().iter().cloned().product();
413        model
414            .wire_node(
415                name,
416                AxisOp::Reshape(
417                    output_shape.h_axis(),
418                    tvec!(geo_collapsed_out.to_dim()),
419                    output_shape.hw_dims().iter().map(|d| d.to_dim()).collect(),
420                ),
421                wire,
422            )
423            .context("in wire_geo_reshape")
424    }
425
426    pub unsafe fn wire_as_lazy_im2col(
427        &self,
428        model: &mut TypedModel,
429        name: &str,
430        wire: &[OutletId],
431    ) -> TractResult<TVec<OutletId>> {
432        let &[mut x, kernel, bias] = wire else { bail!("Wrong number of inputs") };
433        let mut x_fact = model.outlet_fact(x)?.clone();
434        let w_fact = model.outlet_fact(kernel)?.clone();
435        let (geo, m, k, n) = self.compute_geo(&x_fact)?;
436        let (mmm, packing) = self.choose_impl(&x_fact, &w_fact, m, k, &n)?;
437        debug!("{name} as lazy_im2col: m={m} k={k} n={n} {mmm:?}");
438        let input_shape = x_fact.shape.as_concrete().unwrap().to_vec();
439        let mut geo = geo.to_concrete(&input_shape)?.into_owned();
440        let mut input_shape: DataShape = self.pool_spec.data_format.shape(input_shape.into())?;
441        let padding = self.pool_spec.computed_padding(input_shape.hw_dims());
442        if padding.iter().any(|axis| axis.pad_before != 0 || axis.pad_after != 0) {
443            let mut pads = vec![(0, 0); x_fact.rank()];
444            for (ix, ax) in padding.iter().enumerate() {
445                pads[input_shape.h_axis() + ix] = (ax.pad_before, ax.pad_after);
446            }
447            let op = crate::ops::array::Pad {
448                mode: crate::ops::array::PadMode::Constant(
449                    Tensor::zero_scalar_dt(x_fact.datum_type)?.into_arc_tensor(),
450                ),
451                pads,
452            };
453            x = model.wire_node(format!("{name}.pad"), op, &[x])?[0];
454            let valid_pool_spec = PoolSpec { padding: Valid, ..self.pool_spec.clone() };
455            x_fact = model.outlet_fact(x)?.clone();
456            let concrete_shape = x_fact.shape.as_concrete().unwrap();
457            input_shape = valid_pool_spec.data_format.shape(concrete_shape.into())?;
458            geo = valid_pool_spec
459                .compute_geo(&x_fact.shape)?
460                .to_concrete(concrete_shape)?
461                .into_owned();
462        }
463        let c_dt = crate::ops::matmul::output_type(x_fact.datum_type);
464        let c_stride = input_shape.c_stride();
465        let size_of_b = x_fact.datum_type.size_of() as isize;
466        let n_byte_offsets: Vec<isize> =
467            geo.patch.centers_offsets().into_iter().map(|x| x * size_of_b).collect();
468        let k_byte_offsets: Vec<isize> = (0..self.input_channels())
469            .flat_map(|ici| {
470                geo.patch
471                    .standard_layout_data_field
472                    .iter()
473                    .map(move |x| (x + (ici * c_stride) as isize) * size_of_b)
474            })
475            .collect();
476        let (mmm_output_shape, c_axis, h_axis) = self.mmm_output_shape(&geo.output_shape)?;
477        let packer = mmm.packings()[packing]
478            .1
479            .downcast_ref::<PackedFormat>()
480            .with_context(|| {
481                format_err!(
482                    "Quand Im2Col expects regular packed format, got {:?}",
483                    mmm.packings()[packing].1
484                )
485            })?
486            .clone();
487        let params = LazyIm2colParams { packer, n_byte_offsets, k_byte_offsets };
488        let x = model.wire_node(
489            format!("{name}.lazyIm2col"),
490            LazyIm2Col { params: Arc::new(params) },
491            &[x],
492        )?[0];
493
494        let kernel = self.wire_kernel_as_g_o_ihw(model, name, kernel)?[0];
495        let wire = self.wire_mm_weights_bias(
496            model,
497            name,
498            x,
499            kernel,
500            bias,
501            mmm,
502            packing,
503            c_dt,
504            mmm_output_shape.clone().into(),
505            k,
506            c_axis,
507            h_axis,
508        )?;
509
510        let wire = self.wire_remove_group(model, name, &wire, &mmm_output_shape, c_axis)?;
511        let wire = self.wire_rm_n_if_needed(model, name, &wire)?;
512        Self::wire_geo_reshape(model, name, &wire, &geo.output_shape)
513    }
514
515    #[allow(clippy::type_complexity)]
516    fn compute_geo(
517        &self,
518        input_fact: &TypedFact,
519    ) -> TractResult<(PoolGeometry, usize, usize, TDim)> {
520        let geo = self.pool_spec.compute_geo(&input_fact.shape)?;
521
522        trace!("output channels: {:?}", self.output_channels());
523        let m = self.output_channels() / self.group;
524        let k = self.input_channels() * self.pool_spec.kernel_shape.iter().product::<usize>()
525            / self.group;
526        let n: TDim =
527            self.pool_spec.output_shape(&input_fact.shape)?.hw_dims().iter().cloned().product();
528        Ok((geo, m, k, n))
529    }
530
531    fn choose_impl(
532        &self,
533        input_fact: &TypedFact,
534        weight_fact: &TypedFact,
535        m: usize,
536        k: usize,
537        n: &TDim,
538    ) -> TractResult<(Box<dyn MatMatMul>, usize)> {
539        let w_dt = weight_fact.datum_type;
540        let x_dt = input_fact.datum_type;
541
542        let acc = if x_dt.is_float() { x_dt } else { i32::datum_type() };
543        if w_dt.is_opaque() {
544            let bqf = weight_fact
545                .opaque_fact
546                .as_ref()
547                .and_then(|of| of.downcast_ref::<BlockQuantFact>())
548                .unwrap();
549            let weight_type = WeightType::BlockQuant(bqf.format.clone());
550            tract_linalg::ops()
551                .mmm_impls()
552                .iter()
553                .filter(|mmm| mmm.internal_type() == acc)
554                .flat_map(|mmm| {
555                    mmm.packings().iter().enumerate().map(move |(ix, p)| (mmm, ix, &p.0, &p.1))
556                })
557                .filter(|(_, _, pa, pb)| {
558                    pb.precursor() == x_dt.into() && pa.precursor() == weight_type
559                })
560                .map(|(mmm, p, _, _)| (mmm.clone(), p))
561                .min_by_key(|(mmm, _)| {
562                    mmm.quality().cost() as isize * 1000 - (mmm.mr() * mmm.nr()) as isize
563                })
564                .context("Not matmu found")
565        } else {
566            let mmm = tract_linalg::ops()
567                .mmm(acc, Some(m), Some(k), n.to_usize().ok())
568                .context("No matmul found")?;
569            let packing = mmm
570                .packings()
571                .iter()
572                .position(|p| {
573                    p.0.precursor() == w_dt.unquantized().into()
574                        && p.1.precursor() == x_dt.unquantized().into()
575                })
576                .context("No packing found")?;
577            Ok((mmm, packing))
578        }
579    }
580
581    #[allow(clippy::too_many_arguments)]
582    fn wire_mm_weights_bias(
583        &self,
584        model: &mut TypedModel,
585        name: &str,
586        input: OutletId,
587        g_o_ihw: OutletId,
588        bias: OutletId,
589        mmm: Box<dyn MatMatMul>,
590        packing: usize,
591        c_datum_type: DatumType,
592        mmm_output_shape: ShapeFact,
593        k: usize,
594        c_m_axis: usize,
595        c_n_axis: usize,
596    ) -> TractResult<TVec<OutletId>> {
597        ensure!(model.outlet_fact(bias)?.datum_type == mmm.internal_type());
598        let a_pack = &mmm.packings()[packing].0;
599        let packed_ker = self
600            .wire_pack_g_o_ihw(model, name, &**a_pack, g_o_ihw)
601            .context("in kernel_as_packed_as")?;
602        let (mut c_to_a_axis_mapping, mut c_to_b_axis_mapping) = (tvec!(), tvec!());
603
604        c_to_a_axis_mapping.push((c_m_axis - 1, 0)); // Group
605        c_to_b_axis_mapping.push((0, 0)); // Batch
606        c_to_b_axis_mapping.push((c_m_axis - 1, 1)); // Group
607
608        let geo = AddMatMulGeometry {
609            k: k.to_dim(),
610            c_to_a_axis_mapping: MapOutputAxisToInput(c_to_a_axis_mapping),
611            c_to_b_axis_mapping: MapOutputAxisToInput(c_to_b_axis_mapping),
612        };
613        let mut ops: Vec<ProtoFusedSpec> =
614            vec![ProtoFusedSpec::AddMatMul { geo, a: 1, b: 0, packings: vec![(packing, None)] }];
615        let mut wires: TVec<OutletId> = tvec!(input, packed_ker);
616        let bias_fact = model.outlet_fact(bias)?;
617        if bias_fact.konst.is_none() || !bias_fact.konst.as_ref().unwrap().is_all_zero()? {
618            let (fused, bias) = self.wire_bias_as_non_linear(model, name, bias, c_m_axis - 1)?;
619            wires.push(bias);
620            ops.push(fused);
621        }
622        ops.push(ProtoFusedSpec::Store(vec![unsafe {
623            mmm.c_view(Some(c_m_axis), Some(c_n_axis))
624        }]));
625        model.wire_node(
626            format!("{name}.matmatmul"),
627            OptMatMul::new(
628                vec![mmm],
629                ModePicker::Single,
630                c_datum_type.fact(mmm_output_shape),
631                Some(c_m_axis),
632                Some(c_n_axis),
633                ops,
634                packing == 0 && self.group == 1,
635            )?,
636            &wires,
637        )
638    }
639
640    pub fn wire_as_depth_wise(
641        &self,
642        model: &mut TypedModel,
643        name: &str,
644        wire: &[OutletId],
645    ) -> TractResult<OutletId> {
646        let &[x, kernel, mut bias] = wire else { bail!("Wrong number of inputs") };
647        let x_fact = model.outlet_fact(x)?.clone();
648        let x_shape = x_fact.shape.as_concrete().unwrap();
649        let ConcretePoolGeometry { input_shape, patch, output_shape } =
650            self.pool_spec.compute_geo(&x_fact.shape)?.to_concrete(x_shape)?.into_owned();
651        let kernel = self.wire_kernel_as_g_o_ihw(model, name, kernel)?;
652        let c_axis = self.pool_spec.data_format.shape(x_shape)?.c_axis();
653        bias = wire_reshape_bias_for_bin(
654            model,
655            name,
656            bias,
657            x_fact.rank(),
658            c_axis,
659            self.output_channels(),
660        )?[0];
661        let op = DepthWise::new(patch, input_shape, output_shape);
662        Ok(model.wire_node(name, op, &[x, kernel[0], bias])?[0])
663    }
664
665    fn declutter_stride_slice_to_downsample(
666        &self,
667        model: &TypedModel,
668        node: &TypedNode,
669    ) -> TractResult<Option<TypedModelPatch>> {
670        let spatial_rank = self.pool_spec.rank();
671        if let Some(axis) = (0..spatial_rank).find(|&ax| {
672            self.pool_spec.stride(ax) > 1
673                && self.pool_spec.padding.valid_dim(ax, self.pool_spec.stride(ax) == 1)
674                && (self.pool_spec.kernel_shape[ax] == 1
675                    || self.pool_spec.dilation(ax) % self.pool_spec.stride(ax) == 0)
676        }) {
677            let input_fact = model.outlet_fact(node.inputs[0])?;
678            let downsample_factor = self.pool_spec.stride(axis);
679            let mut new_op = self.clone();
680            if new_op.pool_spec.dilation(axis) > 1 {
681                new_op.pool_spec.dilations.as_mut().unwrap()[axis] =
682                    new_op.pool_spec.dilations.as_mut().unwrap()[axis].divceil(downsample_factor);
683            }
684            new_op.pool_spec.strides.as_mut().unwrap()[axis] /= downsample_factor;
685            let mut patch = TypedModelPatch::default();
686            let mut taps = patch.taps(model, &node.inputs)?;
687            let shape = self.pool_spec.data_format.shape(&input_fact.shape)?;
688            taps[0] = patch.wire_node(
689                format!("{}.downsample.{}", node.name, axis),
690                crate::ops::Downsample::new(axis + shape.h_axis(), downsample_factor as isize, 0),
691                &[taps[0]],
692            )?[0];
693            let id = patch.wire_node(&*node.name, new_op, &taps)?[0];
694            patch.shunt_outside(model, OutletId::new(node.id, 0), id)?;
695            return Ok(Some(patch));
696        }
697        Ok(None)
698    }
699
700    fn declutter_as_einsum(
701        &self,
702        model: &TypedModel,
703        node: &TypedNode,
704    ) -> TractResult<Option<TypedModelPatch>> {
705        let (input_facts, output_facts) = model.node_facts(node.id)?;
706        let full_input_shape = input_facts[0].shape.to_tvec();
707        let input_shape = self.pool_spec.data_format.shape(&full_input_shape)?;
708        if self.group == 1
709            && self.pool_spec.strides().iter().all(|s| *s == 1)
710            && self.pool_spec.dilations().iter().all(|d| *d == 1)
711            && self.pool_spec.kernel_shape.iter().product::<usize>() == 1
712            && self
713                .pool_spec
714                .computed_padding(input_shape.hw_dims())
715                .iter()
716                .all(|pad| pad.pad_after.is_zero() && pad.pad_before.is_zero())
717        {
718            let mut axes = self.axes_mapping(&input_facts, &output_facts)?;
719            let mut patch = TypedModelPatch::new("declutter_as_einsum");
720            let mut taps = patch.taps(model, &node.inputs)?;
721            let name = &node.name;
722            let co = self.output_channels();
723            taps[1] =
724                self.wire_kernel_as_g_o_ihw(&mut patch, &format!("{name}.filters"), taps[1])?[0];
725            taps[1] =
726                patch.wire_node(format!("{name}.filters_as_co_ci"), AxisOp::Rm(0), &[taps[1]])?[0];
727
728            while axes.rank(InOut::In(1)) > 0 {
729                axes = axes.remove_axis_occurency(InOut::In(1), 0)?;
730            }
731            axes = axes
732                .with_extra_axis_occurency('O', InOut::In(1), 0)?
733                .with_extra_axis_occurency('I', InOut::In(1), 1)?;
734
735            let bias_fact = input_facts[2];
736            let wire = if self.q_params.is_some() {
737                if bias_fact.rank() == 1 {
738                    axes = axes.linking('O', (InOut::In(2), 0))?;
739                }
740                let op = EinSum { axes, operating_dt: i32::datum_type(), q_params: self.q_params };
741                patch.wire_node(format!("{name}.einsum"), op, &taps)?[0]
742            } else {
743                axes = axes.remove_slot(InOut::In(2))?;
744                let op = EinSum { axes, operating_dt: input_facts[0].datum_type, q_params: None };
745                let mut wire = patch.wire_node(format!("{name}.einsum"), op, &taps[0..2])?[0];
746
747                if !bias_fact.konst.as_ref().map(|f| f.is_zero()).transpose()?.unwrap_or(false) {
748                    let bias_current_shape =
749                        if bias_fact.rank() == 0 { tvec!() } else { tvec!(co.to_dim()) };
750                    let mut bias_shape = tvec!(1.to_dim(); input_shape.rank());
751                    if bias_fact.rank() > 0 {
752                        bias_shape[input_shape.c_axis()] = co.to_dim();
753                    }
754                    let b = patch.wire_node(
755                        format!("{name}.bias.reshape"),
756                        AxisOp::Reshape(0, bias_current_shape, bias_shape),
757                        &[taps[2]],
758                    )?[0];
759                    wire = patch.wire_node(
760                        format!("{name}.bias"),
761                        crate::ops::math::add(),
762                        &[wire, b],
763                    )?[0];
764                }
765                wire
766            };
767            patch.node_mut(wire.node).name = node.name.to_string();
768            patch.shunt_outside(model, node.id.into(), wire)?;
769            return Ok(Some(patch));
770        }
771        Ok(None)
772    }
773
774    fn declutter_precursor_padding(
775        &self,
776        model: &TypedModel,
777        node: &TypedNode,
778    ) -> TractResult<Option<TypedModelPatch>> {
779        rule_if!(!matches!(
780            self.pool_spec.padding,
781            ExplicitOnnxPool(_, _, _) | SameLower | SameUpper
782        ));
783        let prec = model.node(node.inputs[0].node);
784        rule_if_some!(pad = prec.op_as::<Pad>());
785        rule_if_let!(PadMode::Constant(value) = &pad.mode);
786        let shape = self.pool_spec.data_format.shape(&model.outlet_fact(node.inputs[0])?.shape)?;
787        rule_if!(value.is_zero()?);
788        rule_if!(pad.pads[shape.c_axis()] == (0, 0));
789        if self.pool_spec.data_format.has_n() {
790            rule_if!(pad.pads[0] == (0, 0));
791        }
792        let mut before: TVec<usize> = pad.pads[shape.hw_axes()].iter().map(|pair| pair.0).collect();
793        let mut after: TVec<usize> = pad.pads[shape.hw_axes()].iter().map(|pair| pair.1).collect();
794        if let Explicit(bef, aft) = &self.pool_spec.padding {
795            izip!(&mut before, bef).for_each(|(pad, cv)| *pad += cv);
796            izip!(&mut after, aft).for_each(|(pad, cv)| *pad += cv);
797        }
798        let padding = Explicit(before, after);
799        let mut new = self.clone();
800        new.pool_spec.padding = padding;
801        let mut patch = TypedModelPatch::default();
802        let mut wire = patch.taps(model, &node.inputs)?;
803        wire[0] = patch.tap_model(model, prec.inputs[0])?;
804        let wire = patch.wire_node(&node.name, new, &wire)?;
805        patch.shunt_outside(model, node.id.into(), wire[0])?;
806        Ok(Some(patch))
807    }
808
809    fn declutter_channel_arithmetic_succ(
810        &self,
811        model: &TypedModel,
812        node: &TypedNode,
813    ) -> TractResult<Option<TypedModelPatch>> {
814        rule_if!(self.q_params.is_none());
815        rule_if!(self.group == 1);
816        rule_if_let!(&[succ_outlet] = &*node.outputs[0].successors);
817        let succ = model.node(succ_outlet.node);
818        rule_if_some!(bin = succ.op_as::<TypedBinOp>());
819        let other_input = succ.inputs[1 - succ_outlet.slot];
820        let axes_mapping = model.node_axes_mapping(succ.id)?;
821        let input_shape =
822            self.pool_spec.data_format.shape(&model.outlet_fact(node.inputs[0])?.shape)?;
823        let conv_c_axis = input_shape.c_axis();
824        rule_if!(
825            axes_mapping.axis((InOut::In(succ_outlet.slot), conv_c_axis))?.inputs
826                [1 - succ_outlet.slot]
827                .len()
828                == 1
829        );
830        let mut other_expected_shape = tvec!(1.to_dim(); input_shape.rank());
831        other_expected_shape[conv_c_axis] = self.output_channels().to_dim();
832        rule_if!(*other_expected_shape == *model.outlet_fact(other_input)?.shape);
833
834        let mut patch = TypedModelPatch::default();
835        let [input, mut kernel, mut bias] = *patch.taps(model, &node.inputs)? else {
836            panic!("Expect three inputs");
837        };
838        let name = &node.name;
839        let succ_name = &succ.name;
840
841        let operand = patch.tap_model(model, other_input)?;
842
843        let renamed_bias = format!("{name}.{succ_name}.bias");
844        let renamed_kernel = format!("{name}.{succ_name}.kernel");
845        bias = wire_reshape_bias_for_bin(
846            &mut patch,
847            format!("{renamed_bias}.reshape"),
848            bias,
849            1,
850            0,
851            self.output_channels(),
852        )?[0];
853
854        let operand = wire_reshape_bias_for_bin(
855            &mut patch,
856            format!("{renamed_bias}.reshape_operand"),
857            operand,
858            1,
859            0,
860            self.output_channels(),
861        )?[0];
862
863        let operand_fact = patch.outlet_fact(operand)?.shape.to_tvec();
864        let kernel_fact = patch.outlet_fact(kernel)?;
865        let mut operand_shape_for_kernel = tvec!(1.to_dim(); 2 + input_shape.hw_rank());
866        operand_shape_for_kernel[self.kernel_fmt.o_axis(&kernel_fact.shape)] =
867            self.output_channels().to_dim();
868        let operand_for_kernel = patch.wire_node(
869            format!("{renamed_kernel}.reshape_operand"),
870            AxisOp::Reshape(0, operand_fact, operand_shape_for_kernel),
871            &[operand],
872        )?[0];
873
874        if bin.0.is::<Sub>() && succ_outlet.slot == 0 {
875            bias = patch.wire_node(&renamed_bias, sub(), &[bias, operand])?[0];
876        } else if bin.0.is::<Sub>() {
877            bias = patch.wire_node(&renamed_bias, sub(), &[operand, bias])?[0];
878        } else if bin.0.is::<Div>() && succ_outlet.slot == 0 {
879            bias = patch.wire_node(&renamed_bias, div(), &[bias, operand])?[0];
880            kernel = patch.wire_node(&renamed_kernel, div(), &[kernel, operand_for_kernel])?[0];
881        } else if bin.0.is::<Div>() {
882            bias = patch.wire_node(&renamed_bias, div(), &[operand, bias])?[0];
883            kernel = patch.wire_node(&renamed_kernel, div(), &[operand_for_kernel, kernel])?[0];
884        } else if bin.0.is::<Add>() {
885            bias = patch.wire_node(&renamed_bias, add(), &[bias, operand])?[0];
886        } else if bin.0.is::<Mul>() {
887            bias = patch.wire_node(&renamed_bias, mul(), &[bias, operand])?[0];
888            kernel = patch.wire_node(&renamed_kernel, mul(), &[kernel, operand_for_kernel])?[0];
889        } else {
890            return Ok(None);
891        };
892        let wire = patch.wire_node(&node.name, self.clone(), &[input, kernel, bias])?[0];
893        patch.shunt_outside(model, succ_outlet.node.into(), wire)?;
894        Ok(Some(patch))
895    }
896}
897
898impl Op for Conv {
899    fn name(&self) -> StaticName {
900        "Conv".into()
901    }
902
903    fn info(&self) -> TractResult<Vec<String>> {
904        let mut info = self.pool_spec.info();
905        info.push(format!("Kernel {:?} (groups:{})", self.kernel_fmt, self.group));
906        Ok(info)
907    }
908
909    fn validation(&self) -> Validation {
910        Validation::Rounding
911    }
912
913    op_as_typed_op!();
914}
915
916impl EvalOp for Conv {
917    fn is_stateless(&self) -> bool {
918        true
919    }
920
921    fn eval(&self, inputs: TVec<TValue>) -> TractResult<TVec<TValue>> {
922        let mut model = TypedModel::default();
923        let wire: TVec<OutletId> = inputs
924            .iter()
925            .enumerate()
926            .map(|(ix, v)| model.add_source(format!("source.{ix}"), v.datum_type().fact(v.shape())))
927            .collect::<TractResult<_>>()?;
928        let wire = unsafe {
929            if self.q_params.is_some() {
930                self.wire_as_quant_im2col(&mut model, "im2col-adhoc", &wire)?
931            } else {
932                self.wire_as_im2col_pair(&mut model, "im2col-adhoc", &wire)?
933            }
934        };
935        model.set_output_outlets(&wire)?;
936        model.into_runnable()?.run(inputs)
937    }
938}
939
940impl TypedOp for Conv {
941    fn output_facts(&self, inputs: &[&TypedFact]) -> TractResult<TVec<TypedFact>> {
942        ensure!(self.q_params.is_some() || inputs[0].datum_type.is_float());
943        let q_inputs = if self.q_params.is_some() { 6 } else { 0 };
944        ensure!(inputs[1].datum_type.is_number() || self.kernel_fmt == KernelFormat::OIHW);
945        if inputs.len() != 3 + q_inputs {
946            bail!("Wrong number of inputs: expected {} got {}", 3 + q_inputs, inputs.len());
947        }
948        if self.q_params.is_some() {
949            ensure!(inputs[2].datum_type == i32::datum_type());
950            ensure!(inputs[3].datum_type == i32::datum_type());
951            ensure!(inputs[4].datum_type.is_float());
952            ensure!(inputs[5].datum_type == i32::datum_type());
953            ensure!(inputs[6].datum_type.is_float());
954            ensure!(inputs[7].datum_type == i32::datum_type());
955            ensure!(inputs[8].datum_type.is_float());
956        }
957        let weight_shape = block_quant_aware_weight_shape(inputs[1])?;
958        ensure!(self.pool_spec.rank() + 2 == weight_shape.len());
959        if self.pool_spec.data_format.shape(&*inputs[0].shape)?.c()
960            != &self.input_channels().to_dim()
961        {
962            bail!(
963                "Inconsistent convolution: input is {:?}, but kernel expects {} input channels.\n{:?}",
964                inputs[0],
965                self.input_channels(),
966                self
967            );
968        }
969        if let ExplicitOnnxPool(bef, after, _) | Explicit(bef, after) = &self.pool_spec.padding {
970            anyhow::ensure!(bef.len() == self.pool_spec.rank());
971            anyhow::ensure!(after.len() == self.pool_spec.rank());
972        }
973        ensure!(
974            inputs[2].rank() == 0
975                || (inputs[2].rank() == 1
976                    && inputs[2].shape.volume() == self.output_channels().to_dim()),
977            "Bias should be scalar or a vector with one value per output channel. Output channels is {}, bias is {:?}",
978            self.output_channels(),
979            inputs[2]
980        );
981        let mut fact = self.pool_spec.output_facts(inputs)?.remove(0);
982        if let Some(dt) = self.q_params {
983            fact.datum_type = dt;
984        } else {
985            ensure!(
986                inputs[1].datum_type.is_opaque() || inputs[0].datum_type == inputs[1].datum_type,
987                "Convolution input, weights and bias must have the same type, got {inputs:?}",
988            )
989        }
990        Ok(tvec!(fact))
991    }
992
993    fn axes_mapping(
994        &self,
995        inputs: &[&TypedFact],
996        outputs: &[&TypedFact],
997    ) -> TractResult<AxesMapping> {
998        let fact = &inputs[0];
999        let shape = self.pool_spec.data_format.shape(&fact.shape)?;
1000        let mut axes = AxesMapping::disconnected(inputs, outputs)?
1001            .renaming((InOut::In(0), shape.c_axis()), 'I')?
1002            .renaming((InOut::Out(0), shape.c_axis()), 'O')?;
1003        if let Some(n_axis) = shape.n_axis() {
1004            axes = axes
1005                .renaming((InOut::In(0), n_axis), 'N')?
1006                .linking('N', (InOut::Out(0), n_axis))?;
1007        }
1008        let h_axis = shape.h_axis();
1009        let geo = "HWXYZ".chars().chain('a'..);
1010        let kernel_spatial_shape = &self.pool_spec.kernel_shape;
1011        let padding = self.pool_spec.computed_padding(shape.hw_dims());
1012        for ((ix, &dim), repr) in kernel_spatial_shape.iter().enumerate().zip(geo) {
1013            if dim == 1
1014                && self.pool_spec.dilation(ix) == 1
1015                && self.pool_spec.stride(ix) == 1
1016                && padding[ix].pad_before.is_zero()
1017                && padding[ix].pad_after.is_zero()
1018            {
1019                axes = axes
1020                    .renaming((InOut::In(0), ix + h_axis), repr)?
1021                    .linking(repr, (InOut::Out(0), ix + h_axis))?;
1022            }
1023        }
1024        if self.q_params.is_some() {
1025            for (qp_ix, qp) in inputs.iter().enumerate().skip(3) {
1026                if qp.rank() == 1 {
1027                    axes = match qp_ix {
1028                        3 | 4 => axes.linking('I', (InOut::In(qp_ix), 0))?,
1029                        5 | 6 => axes.linking('O', (InOut::In(qp_ix), 0))?,
1030                        7 | 8 => axes.linking('O', (InOut::In(qp_ix), 0))?,
1031                        _ => unreachable!(),
1032                    };
1033                }
1034            }
1035        }
1036        Ok(axes)
1037    }
1038
1039    fn declutter(
1040        &self,
1041        model: &TypedModel,
1042        node: &TypedNode,
1043    ) -> TractResult<Option<TypedModelPatch>> {
1044        macro_rules! pass {
1045            ($func:ident) => {
1046                if let Some(mut r) = self.$func(model, node).context(stringify!($func))? {
1047                    trace!(stringify!($func));
1048                    r.push_context(stringify!($func));
1049                    return Ok(Some(r));
1050                }
1051            };
1052        }
1053        pass!(declutter_stride_slice_to_downsample);
1054        pass!(declutter_as_einsum);
1055        pass!(declutter_channel_arithmetic_succ);
1056        pass!(declutter_precursor_padding);
1057        Ok(None)
1058    }
1059
1060    fn cost(&self, inputs: &[&TypedFact]) -> TractResult<TVec<(Cost, TDim)>> {
1061        let shape = self.pool_spec.data_format.shape(inputs[0].shape.to_tvec())?;
1062        let kernel_spatial_shape = &self.pool_spec.kernel_shape;
1063        let output_dims = self.pool_spec.padding.compute(
1064            shape.hw_dims(),
1065            kernel_spatial_shape,
1066            &self
1067                .pool_spec
1068                .dilations
1069                .clone()
1070                .unwrap_or_else(|| tvec!(1; kernel_spatial_shape.len())),
1071            &self.pool_spec.strides.clone().unwrap_or_else(|| tvec!(1; kernel_spatial_shape.len())),
1072        );
1073        let n_output_points: TDim =
1074            output_dims.iter().map(|d| d.convoluted.clone()).product::<TDim>();
1075        let n_output_channels = self.output_channels().to_dim();
1076        let kernel_surface = kernel_spatial_shape.iter().product::<usize>().to_dim();
1077        let one = 1.to_dim();
1078        Ok(tvec!((
1079            Cost::FMA(inputs[0].datum_type),
1080            shape.n().cloned().unwrap_or(one)
1081                * shape.c()
1082                * n_output_channels
1083                * n_output_points
1084                * kernel_surface
1085                / self.group
1086        )))
1087    }
1088
1089    fn change_axes(
1090        &self,
1091        model: &TypedModel,
1092        node: &TypedNode,
1093        io: InOut,
1094        change: &AxisOp,
1095    ) -> TractResult<Option<AxisChangeConsequence>> {
1096        if io == InOut::In(1) {
1097            return Ok(None);
1098        }
1099        if io == InOut::In(2) {
1100            if let &AxisOp::Rm(_) = change {
1101                return Ok(Some(AxisChangeConsequence {
1102                    substitute_op: Some(Box::new(self.clone())),
1103                    wire_changes: tvec!(),
1104                }));
1105            }
1106        }
1107        let full_input_shape = model.outlet_fact(node.inputs[0])?.shape.to_tvec();
1108        let shape = self.pool_spec.data_format.shape(full_input_shape.clone())?;
1109        // remove n
1110        if let Some(n) = shape.n_axis() {
1111            assert_eq!(n, 0);
1112            if change == &AxisOp::Rm(n) {
1113                let op = Conv { pool_spec: self.pool_spec.dispose_n_axis(), ..self.clone() };
1114                return Ok(Some(AxisChangeConsequence {
1115                    substitute_op: Some(Box::new(op)),
1116                    wire_changes: tvec!(
1117                        (InOut::In(0), change.clone()),
1118                        (InOut::Out(0), change.clone())
1119                    ),
1120                }));
1121            }
1122            if change.transform_axis(n).map(|axis| axis > 0).unwrap_or(true) {
1123                return Ok(None);
1124            }
1125        }
1126        // format swap: chw <-> hwc
1127        let (new_format, axis_move) = match self.pool_spec.data_format {
1128            DataFormat::NCHW => {
1129                (DataFormat::NHWC, AxisOp::Move(shape.c_axis(), full_input_shape.len() - 1))
1130            }
1131            DataFormat::CHW => {
1132                (DataFormat::HWC, AxisOp::Move(shape.c_axis(), full_input_shape.len() - 1))
1133            }
1134            DataFormat::NHWC => (DataFormat::NCHW, AxisOp::Move(shape.c_axis(), 1)),
1135            DataFormat::HWC => (DataFormat::CHW, AxisOp::Move(shape.c_axis(), 0)),
1136        };
1137        if *change == axis_move {
1138            let mut new_op = self.clone();
1139            new_op.pool_spec.data_format = new_format;
1140            return Ok(Some(AxisChangeConsequence {
1141                substitute_op: Some(Box::new(new_op)),
1142                wire_changes: tvec!(
1143                    (InOut::In(0), change.clone()),
1144                    (InOut::Out(0), change.clone())
1145                ),
1146            }));
1147        }
1148        // geo axis manips
1149        if model.node_input_facts(node.id)?[1].datum_type.is_opaque() {
1150            return Ok(None);
1151        }
1152        use AxisOp::*;
1153        let h_axis = shape.h_axis();
1154        let hw_axes = shape.hw_axes();
1155        let kh_axis = self.kernel_fmt.h_axis();
1156        let (geo_adjusted, kernel_adjusted) = match change {
1157            Rm(a)
1158                if hw_axes.contains(a)
1159                    && hw_axes.len() > 1
1160                    && self.pool_spec.dilation(a - h_axis) == 1
1161                    && self.pool_spec.stride(a - h_axis) == 1
1162                    && self.pool_spec.kernel_shape[a - h_axis] == 1 =>
1163            {
1164                let geo_axis = a - h_axis;
1165                (Rm(geo_axis), Rm(kh_axis + geo_axis))
1166            }
1167            Add(a) if hw_axes.contains(a) => (Add(a - h_axis), Add(a - h_axis + kh_axis)),
1168            Move(f, t) if hw_axes.contains(f) && hw_axes.contains(t) => {
1169                (Move(f - h_axis, t - h_axis), Move(f - h_axis + kh_axis, t - h_axis + kh_axis))
1170            }
1171            _ => return Ok(None),
1172        };
1173        let pool_spec = self.pool_spec.change_geo_axes(&geo_adjusted)?;
1174        let new_op = Conv { pool_spec, ..self.clone() };
1175        Ok(Some(AxisChangeConsequence {
1176            substitute_op: Some(Box::new(new_op)),
1177            wire_changes: tvec!(
1178                (InOut::In(0), change.clone()),
1179                (InOut::In(1), kernel_adjusted),
1180                (InOut::Out(0), change.clone())
1181            ),
1182        }))
1183    }
1184
1185    fn codegen(
1186        &self,
1187        model: &TypedModel,
1188        node: &TypedNode,
1189    ) -> TractResult<Option<TypedModelPatch>> {
1190        let input_fact = model.outlet_fact(node.inputs[0])?;
1191        unsafe {
1192            if self.q_params.is_some() {
1193                let mut patch = TypedModelPatch::new("quantized-codegen");
1194                let inputs = patch.taps(model, &node.inputs)?;
1195                let wire = self
1196                    .wire_as_quant_im2col(&mut patch, &node.name, &inputs)
1197                    .context("in wire_as_quant_im2col")?;
1198                patch.shunt_outside(model, node.id.into(), wire[0])?;
1199                patch.obliterate(node.id)?;
1200                Ok(Some(patch))
1201            } else if input_fact
1202                .shape
1203                .as_concrete()
1204                .map(|s| {
1205                    should_use_lazy(
1206                        &self.pool_spec.data_format.shape(s.into()).unwrap(),
1207                        &self.pool_spec,
1208                        self.group,
1209                    )
1210                })
1211                .unwrap_or(false)
1212            {
1213                let mut patch = TypedModelPatch::new("lazy-im2col");
1214                let inputs = patch.taps(model, &node.inputs)?;
1215                let wire = self
1216                    .wire_as_lazy_im2col(&mut patch, &node.name, &inputs)
1217                    .context("wire_as_lazy_im2col")?[0];
1218                patch.shunt_outside(model, OutletId::new(node.id, 0), wire)?;
1219                patch.obliterate(node.id)?;
1220                Ok(Some(patch))
1221            } else if self.group != 1
1222                && self.group == self.output_channels()
1223                && self.group == self.input_channels()
1224                && input_fact.shape.as_concrete().is_some()
1225            {
1226                let mut patch = TypedModelPatch::new("depth_wise");
1227                let inputs = patch.taps(model, &node.inputs)?;
1228                let wire = self
1229                    .wire_as_depth_wise(&mut patch, &node.name, &inputs)
1230                    .context("wire_as_depth_wise")?;
1231                patch.shunt_outside(model, OutletId::new(node.id, 0), wire)?;
1232                patch.obliterate(node.id)?;
1233                Ok(Some(patch))
1234            } else {
1235                let mut patch = TypedModelPatch::new("im2col");
1236                let inputs = patch.taps(model, &node.inputs)?;
1237                let wire = self
1238                    .wire_as_im2col_pair(&mut patch, &node.name, &inputs)
1239                    .context("in wire_as_im2col_pair")?[0];
1240                patch.shunt_outside(model, OutletId::new(node.id, 0), wire)?;
1241                patch.obliterate(node.id)?;
1242                Ok(Some(patch))
1243            }
1244        }
1245    }
1246
1247    as_op!();
1248}
1249
1250fn should_use_lazy(input_shape: &DataShape, pool_spec: &PoolSpec, group: usize) -> bool {
1251    input_shape.n().unwrap_or(&1) == &1
1252        && group == 1
1253        && pool_spec.kernel_shape.iter().product::<usize>() > 5
1254}
1255
1256#[allow(non_snake_case)]
1257#[cfg(test)]
1258mod test {
1259    use super::*;
1260    use crate::ops::array::Pad;
1261    use DataFormat::*;
1262
1263    #[test]
1264    fn onnx_basic_convinteger() {
1265        let op = Conv {
1266            pool_spec: PoolSpec {
1267                data_format: NCHW,
1268                kernel_shape: tvec!(2, 2),
1269                padding: Valid,
1270                dilations: None,
1271                strides: None,
1272                input_channels: 1,
1273                output_channels: 1,
1274            },
1275            kernel_fmt: KernelFormat::OIHW,
1276            group: 1,
1277            q_params: Some(i32::datum_type()),
1278        };
1279        let input = tvec!(
1280            rctensor4(&[[[[1u8, 2, 3], [4, 5, 6], [7, 8, 9]]]]),
1281            rctensor4(&[[[[1u8, 1], [1, 1]]]]),
1282            rctensor0(0u32),
1283            rctensor0(1u8),
1284            rctensor0(1.0f32),
1285            rctensor0(0u8),
1286            rctensor0(1.0f32),
1287            rctensor0(0i32),
1288            rctensor0(1.0f32),
1289        );
1290        let input = input.into_iter().map(IntoTValue::into_tvalue).collect::<TVec<_>>();
1291        let output = op.eval(input).unwrap();
1292        assert_eq!(*output[0], tensor4(&[[[[8i32, 12], [20, 24]]]]));
1293    }
1294
1295    #[test]
1296    fn valid_conv_absorbs_precursor_pad() -> TractResult<()> {
1297        let mut model = TypedModel::default();
1298        let wire = tvec!(model.add_source("source", f32::fact(dims!(1, 10)))?);
1299        let wire = model.wire_node(
1300            "pad",
1301            Pad {
1302                pads: vec![(0, 0), (1, 0)],
1303                mode: ops::array::PadMode::Constant(rctensor0(0f32)),
1304            },
1305            &wire,
1306        )?;
1307        let kernel = model.add_const("kernel", rctensor3(&[[[1f32, 2f32]]]))?;
1308        let bias = model.add_const("bias", rctensor0(0f32))?;
1309        let wire = model.wire_node(
1310            "conv",
1311            Conv {
1312                pool_spec: PoolSpec {
1313                    data_format: crate::ops::nn::DataFormat::CHW,
1314                    dilations: None,
1315                    strides: None,
1316                    kernel_shape: tvec![2],
1317                    padding: Explicit(tvec![0], tvec![0]),
1318                    input_channels: 1,
1319                    output_channels: 1,
1320                },
1321                kernel_fmt: crate::ops::cnn::KernelFormat::OIHW,
1322                group: 1,
1323                q_params: None,
1324            },
1325            &[wire[0], kernel, bias],
1326        )?;
1327        model.set_output_outlets(&wire)?;
1328        model.declutter()?;
1329        assert_eq!(model.nodes().len(), 4); // source + conv + kernel + bias
1330        let cv = model.nodes()[3].op_as::<Conv>().unwrap();
1331        assert_eq!(cv.pool_spec.padding, Explicit(tvec![1], tvec![0])); // source + conv
1332        Ok(())
1333    }
1334}