use core::arch::x86_64::*;
use crate::compat::*;
use crate::convolution::optimisations::Normalizer16;
use crate::pixels::U8;
use crate::{simd_utils, ImageView, ImageViewMut};
#[inline]
pub(crate) fn horiz_convolution(
src_view: &impl ImageView<Pixel = U8>,
dst_view: &mut impl ImageViewMut<Pixel = U8>,
offset: u32,
normalizer: &Normalizer16,
) {
let dst_height = dst_view.height();
let src_iter = src_view.iter_4_rows(offset, dst_height + offset);
let dst_iter = dst_view.iter_4_rows_mut();
for (src_rows, dst_rows) in src_iter.zip(dst_iter) {
unsafe {
horiz_convolution_four_rows(src_rows, dst_rows, normalizer);
}
}
let yy = dst_height - dst_height % 4;
let src_rows = src_view.iter_rows(yy + offset);
let dst_rows = dst_view.iter_rows_mut(yy);
for (src_row, dst_row) in src_rows.zip(dst_rows) {
unsafe {
horiz_convolution_one_row(src_row, dst_row, normalizer);
}
}
}
#[inline]
#[target_feature(enable = "avx2")]
unsafe fn horiz_convolution_four_rows(
src_rows: [&[U8]; 4],
dst_rows: [&mut [U8]; 4],
normalizer: &Normalizer16,
) {
let zero = _mm_setzero_si128();
let initial = _mm256_set1_epi32(1 << (normalizer.precision() - 4));
for (dst_x, chunk) in normalizer.chunks().iter().enumerate() {
let mut x = chunk.start as usize;
let mut result_i32x8x4 = [initial, initial, initial, initial];
let coeffs_by_16 = chunk.values().chunks_exact(16);
let reminder16 = coeffs_by_16.remainder();
for k in coeffs_by_16 {
let coeffs_i16x16 = _mm256_loadu_si256(k.as_ptr() as *const __m256i);
for i in 0..4 {
let pixels_u8x16 = simd_utils::loadu_si128(src_rows[i], x);
let pixels_i16x16 = _mm256_cvtepu8_epi16(pixels_u8x16);
result_i32x8x4[i] = _mm256_add_epi32(
result_i32x8x4[i],
_mm256_madd_epi16(pixels_i16x16, coeffs_i16x16),
);
}
x += 16;
}
let mut coeffs_by_8 = reminder16.chunks_exact(8);
let reminder8 = coeffs_by_8.remainder();
if let Some(k) = coeffs_by_8.next() {
let coeffs_i16x8 = _mm_loadu_si128(k.as_ptr() as *const __m128i);
for i in 0..4 {
let pixels_u8x8 = simd_utils::loadl_epi64(src_rows[i], x);
let pixels_i16x8 = _mm_cvtepu8_epi16(pixels_u8x8);
result_i32x8x4[i] = _mm256_add_epi32(
result_i32x8x4[i],
_mm256_set_m128i(zero, _mm_madd_epi16(pixels_i16x8, coeffs_i16x8)),
);
}
x += 8;
}
let mut result_i32x4 = result_i32x8x4.map(|v| hsum_i32x8_avx2(v));
for &coeff in reminder8 {
let coeff_i32 = coeff as i32;
for i in 0..4 {
result_i32x4[i] += src_rows[i].get_unchecked(x).0.to_owned() as i32 * coeff_i32;
}
x += 1;
}
let result_u8x4 = result_i32x4.map(|v| normalizer.clip(v));
for i in 0..4 {
dst_rows[i].get_unchecked_mut(dst_x).0 = result_u8x4[i];
}
}
}
#[inline]
#[target_feature(enable = "avx2")]
unsafe fn horiz_convolution_one_row(src_row: &[U8], dst_row: &mut [U8], normalizer: &Normalizer16) {
let zero = _mm_setzero_si128();
let initial = _mm256_set1_epi32(1 << (normalizer.precision() - 4));
for (dst_x, chunk) in normalizer.chunks().iter().enumerate() {
let mut x = chunk.start as usize;
let mut result_i32x8 = initial;
let coeffs_by_16 = chunk.values().chunks_exact(16);
let reminder16 = coeffs_by_16.remainder();
for k in coeffs_by_16 {
let coeffs_i16x16 = _mm256_loadu_si256(k.as_ptr() as *const __m256i);
let pixels_u8x16 = simd_utils::loadu_si128(src_row, x);
let pixels_i16x16 = _mm256_cvtepu8_epi16(pixels_u8x16);
result_i32x8 = _mm256_add_epi32(
result_i32x8,
_mm256_madd_epi16(pixels_i16x16, coeffs_i16x16),
);
x += 16;
}
let mut coeffs_by_8 = reminder16.chunks_exact(8);
let reminder8 = coeffs_by_8.remainder();
if let Some(k) = coeffs_by_8.next() {
let coeffs_i16x8 = _mm_loadu_si128(k.as_ptr() as *const __m128i);
let pixels_u8x8 = simd_utils::loadl_epi64(src_row, x);
let pixels_i16x8 = _mm_cvtepu8_epi16(pixels_u8x8);
result_i32x8 = _mm256_add_epi32(
result_i32x8,
_mm256_set_m128i(zero, _mm_madd_epi16(pixels_i16x8, coeffs_i16x8)),
);
x += 8;
}
let mut result_i32 = hsum_i32x8_avx2(result_i32x8);
for &coeff in reminder8 {
let coeff_i32 = coeff as i32;
result_i32 += src_row.get_unchecked(x).0 as i32 * coeff_i32;
x += 1;
}
dst_row.get_unchecked_mut(dst_x).0 = normalizer.clip(result_i32);
}
}
#[inline]
#[target_feature(enable = "avx2")]
unsafe fn hsum_i32x8_avx2(v: __m256i) -> i32 {
let sum128 = _mm_add_epi32(_mm256_castsi256_si128(v), _mm256_extracti128_si256::<1>(v));
hsum_epi32_avx(sum128)
}
#[inline]
#[target_feature(enable = "avx2")]
unsafe fn hsum_epi32_avx(x: __m128i) -> i32 {
let hi64 = _mm_unpackhi_epi64(x, x);
let sum64 = _mm_add_epi32(hi64, x);
const I: i32 = (2 << 6) | (3 << 4) | 1;
let hi32 = _mm_shuffle_epi32::<I>(sum64); let sum32 = _mm_add_epi32(sum64, hi32);
_mm_cvtsi128_si32(sum32) }