use core::arch::x86_64::*;
use crate::convolution::optimisations::Normalizer32;
use crate::pixels::U16;
use crate::{simd_utils, ImageView, ImageViewMut};
#[inline]
pub(crate) fn horiz_convolution(
src_view: &impl ImageView<Pixel = U16>,
dst_view: &mut impl ImageViewMut<Pixel = U16>,
offset: u32,
normalizer: &Normalizer32,
) {
let dst_height = dst_view.height();
let src_iter = src_view.iter_4_rows(offset, dst_height + offset);
let dst_iter = dst_view.iter_4_rows_mut();
for (src_rows, dst_rows) in src_iter.zip(dst_iter) {
unsafe {
horiz_convolution_four_rows(src_rows, dst_rows, normalizer);
}
}
let yy = dst_height - dst_height % 4;
let src_rows = src_view.iter_rows(yy + offset);
let dst_rows = dst_view.iter_rows_mut(yy);
for (src_row, dst_row) in src_rows.zip(dst_rows) {
unsafe {
horiz_convolution_one_row(src_row, dst_row, normalizer);
}
}
}
#[target_feature(enable = "avx2")]
unsafe fn horiz_convolution_four_rows(
src_rows: [&[U16]; 4],
dst_rows: [&mut [U16]; 4],
normalizer: &Normalizer32,
) {
let precision = normalizer.precision();
let half_error = 1i64 << (precision - 1);
let mut ll_buf = [0i64; 4];
#[rustfmt::skip]
let l0l1_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 3, 2, -1, -1, -1, -1, -1, -1, 1, 0,
-1, -1, -1, -1, -1, -1, 3, 2, -1, -1, -1, -1, -1, -1, 1, 0,
);
#[rustfmt::skip]
let l2l3_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 7, 6, -1, -1, -1, -1, -1, -1, 5, 4,
-1, -1, -1, -1, -1, -1, 7, 6, -1, -1, -1, -1, -1, -1, 5, 4,
);
#[rustfmt::skip]
let l4l5_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 11, 10, -1, -1, -1, -1, -1, -1, 9, 8,
-1, -1, -1, -1, -1, -1, 11, 10, -1, -1, -1, -1, -1, -1, 9, 8,
);
#[rustfmt::skip]
let l6l7_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 15, 14, -1, -1, -1, -1, -1, -1, 13, 12,
-1, -1, -1, -1, -1, -1, 15, 14, -1, -1, -1, -1, -1, -1, 13, 12,
);
for (dst_x, chunk) in normalizer.chunks().iter().enumerate() {
let mut x = chunk.start as usize;
let mut ll_sum = [_mm256_set1_epi64x(0); 2];
let mut coeffs = chunk.values();
let coeffs_by_8 = coeffs.chunks_exact(8);
coeffs = coeffs_by_8.remainder();
for k in coeffs_by_8 {
let coeff01_i64x4 =
_mm256_set_epi64x(k[1] as i64, k[0] as i64, k[1] as i64, k[0] as i64);
let coeff23_i64x4 =
_mm256_set_epi64x(k[3] as i64, k[2] as i64, k[3] as i64, k[2] as i64);
let coeff45_i64x4 =
_mm256_set_epi64x(k[5] as i64, k[4] as i64, k[5] as i64, k[4] as i64);
let coeff67_i64x4 =
_mm256_set_epi64x(k[7] as i64, k[6] as i64, k[7] as i64, k[6] as i64);
for (i, sum) in ll_sum.iter_mut().enumerate() {
let source = _mm256_set_m128i(
simd_utils::loadu_si128(src_rows[i * 2 + 1], x),
simd_utils::loadu_si128(src_rows[i * 2], x),
);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l0l1_i64x4, coeff01_i64x4));
let l2l3_i64x4 = _mm256_shuffle_epi8(source, l2l3_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l2l3_i64x4, coeff23_i64x4));
let l4l5_i64x4 = _mm256_shuffle_epi8(source, l4l5_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l4l5_i64x4, coeff45_i64x4));
let l6l7_i64x4 = _mm256_shuffle_epi8(source, l6l7_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l6l7_i64x4, coeff67_i64x4));
}
x += 8;
}
let coeffs_by_4 = coeffs.chunks_exact(4);
coeffs = coeffs_by_4.remainder();
for k in coeffs_by_4 {
let coeff01_i64x4 =
_mm256_set_epi64x(k[1] as i64, k[0] as i64, k[1] as i64, k[0] as i64);
let coeff23_i64x4 =
_mm256_set_epi64x(k[3] as i64, k[2] as i64, k[3] as i64, k[2] as i64);
for (i, sum) in ll_sum.iter_mut().enumerate() {
let source = _mm256_set_m128i(
simd_utils::loadl_epi64(src_rows[i * 2 + 1], x),
simd_utils::loadl_epi64(src_rows[i * 2], x),
);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l0l1_i64x4, coeff01_i64x4));
let l2l3_i64x4 = _mm256_shuffle_epi8(source, l2l3_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l2l3_i64x4, coeff23_i64x4));
}
x += 4;
}
let coeffs_by_2 = coeffs.chunks_exact(2);
coeffs = coeffs_by_2.remainder();
for k in coeffs_by_2 {
let coeff01_i64x4 =
_mm256_set_epi64x(k[1] as i64, k[0] as i64, k[1] as i64, k[0] as i64);
for (i, sum) in ll_sum.iter_mut().enumerate() {
let source = _mm256_set_m128i(
simd_utils::loadl_epi32(src_rows[i * 2 + 1], x),
simd_utils::loadl_epi32(src_rows[i * 2], x),
);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(l0l1_i64x4, coeff01_i64x4));
}
x += 2;
}
for &k in coeffs {
let coeff0_i64x4 = _mm256_set1_epi64x(k as i64);
for (i, sum) in ll_sum.iter_mut().enumerate() {
let source = _mm256_set_epi64x(
0,
src_rows[i * 2 + 1].get_unchecked(x).0 as i64,
0,
src_rows[i * 2].get_unchecked(x).0 as i64,
);
*sum = _mm256_add_epi64(*sum, _mm256_mul_epi32(source, coeff0_i64x4));
}
}
for (i, &ll) in ll_sum.iter().enumerate() {
_mm256_storeu_si256(ll_buf.as_mut_ptr() as *mut __m256i, ll);
let dst_pixel = dst_rows[i * 2].get_unchecked_mut(dst_x);
dst_pixel.0 = normalizer.clip(ll_buf[0] + ll_buf[1] + half_error);
let dst_pixel = dst_rows[i * 2 + 1].get_unchecked_mut(dst_x);
dst_pixel.0 = normalizer.clip(ll_buf[2] + ll_buf[3] + half_error);
}
}
}
#[target_feature(enable = "avx2")]
unsafe fn horiz_convolution_one_row(
src_row: &[U16],
dst_row: &mut [U16],
normalizer: &Normalizer32,
) {
let precision = normalizer.precision();
let half_error = 1i64 << (precision - 1);
let mut ll_buf = [0i64; 4];
#[rustfmt::skip]
let l0l1_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 3, 2, -1, -1, -1, -1, -1, -1, 1, 0,
-1, -1, -1, -1, -1, -1, 3, 2, -1, -1, -1, -1, -1, -1, 1, 0,
);
#[rustfmt::skip]
let l2l3_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 7, 6, -1, -1, -1, -1, -1, -1, 5, 4,
-1, -1, -1, -1, -1, -1, 7, 6, -1, -1, -1, -1, -1, -1, 5, 4,
);
#[rustfmt::skip]
let l4l5_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 11, 10, -1, -1, -1, -1, -1, -1, 9, 8,
-1, -1, -1, -1, -1, -1, 11, 10, -1, -1, -1, -1, -1, -1, 9, 8,
);
#[rustfmt::skip]
let l6l7_shuffle = _mm256_set_epi8(
-1, -1, -1, -1, -1, -1, 15, 14, -1, -1, -1, -1, -1, -1, 13, 12,
-1, -1, -1, -1, -1, -1, 15, 14, -1, -1, -1, -1, -1, -1, 13, 12,
);
for (dst_x, chunk) in normalizer.chunks().iter().enumerate() {
let mut x = chunk.start as usize;
let mut ll_sum = _mm256_set1_epi64x(0);
let mut coeffs = chunk.values();
let coeffs_by_16 = coeffs.chunks_exact(16);
coeffs = coeffs_by_16.remainder();
for k in coeffs_by_16 {
let coeff0189_i64x4 =
_mm256_set_epi64x(k[9] as i64, k[8] as i64, k[1] as i64, k[0] as i64);
let coeff23ab_i64x4 =
_mm256_set_epi64x(k[11] as i64, k[10] as i64, k[3] as i64, k[2] as i64);
let coeff45cd_i64x4 =
_mm256_set_epi64x(k[13] as i64, k[12] as i64, k[5] as i64, k[4] as i64);
let coeff67ef_i64x4 =
_mm256_set_epi64x(k[15] as i64, k[14] as i64, k[7] as i64, k[6] as i64);
let source = simd_utils::loadu_si256(src_row, x);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l0l1_i64x4, coeff0189_i64x4));
let l2l3_i64x4 = _mm256_shuffle_epi8(source, l2l3_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l2l3_i64x4, coeff23ab_i64x4));
let l4l5_i64x4 = _mm256_shuffle_epi8(source, l4l5_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l4l5_i64x4, coeff45cd_i64x4));
let l6l7_i64x4 = _mm256_shuffle_epi8(source, l6l7_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l6l7_i64x4, coeff67ef_i64x4));
x += 16;
}
let coeffs_by_8 = coeffs.chunks_exact(8);
coeffs = coeffs_by_8.remainder();
for k in coeffs_by_8 {
let coeff0145_i64x4 =
_mm256_set_epi64x(k[5] as i64, k[4] as i64, k[1] as i64, k[0] as i64);
let coeff2367_i64x4 =
_mm256_set_epi64x(k[7] as i64, k[6] as i64, k[3] as i64, k[2] as i64);
let source = _mm256_set_m128i(
simd_utils::loadl_epi64(src_row, x + 4),
simd_utils::loadl_epi64(src_row, x),
);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l0l1_i64x4, coeff0145_i64x4));
let l2l3_i64x4 = _mm256_shuffle_epi8(source, l2l3_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l2l3_i64x4, coeff2367_i64x4));
x += 8;
}
let coeffs_by_4 = coeffs.chunks_exact(4);
coeffs = coeffs_by_4.remainder();
for k in coeffs_by_4 {
let coeff0123_i64x4 =
_mm256_set_epi64x(k[3] as i64, k[2] as i64, k[1] as i64, k[0] as i64);
let source = _mm256_set_m128i(
simd_utils::loadl_epi32(src_row, x + 2),
simd_utils::loadl_epi32(src_row, x),
);
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l0l1_i64x4, coeff0123_i64x4));
x += 4;
}
let coeffs_by_2 = coeffs.chunks_exact(2);
coeffs = coeffs_by_2.remainder();
for k in coeffs_by_2 {
let coeff01_i64x4 = _mm256_set_epi64x(0, 0, k[1] as i64, k[0] as i64);
let source = _mm256_set_m128i(_mm_setzero_si128(), simd_utils::loadl_epi32(src_row, x));
let l0l1_i64x4 = _mm256_shuffle_epi8(source, l0l1_shuffle);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(l0l1_i64x4, coeff01_i64x4));
x += 2;
}
for &k in coeffs {
let coeff0_i64x4 = _mm256_set1_epi64x(k as i64);
let source = _mm256_set_epi64x(0, 0, 0, src_row.get_unchecked(x).0 as i64);
ll_sum = _mm256_add_epi64(ll_sum, _mm256_mul_epi32(source, coeff0_i64x4));
}
_mm256_storeu_si256(ll_buf.as_mut_ptr() as *mut __m256i, ll_sum);
let dst_pixel = dst_row.get_unchecked_mut(dst_x);
dst_pixel.0 = normalizer.clip(ll_buf.iter().sum::<i64>() + half_error);
}
}