use core::arch::aarch64::*;
use super::native;
use crate::pixels::U8x2;
use crate::utils::foreach_with_pre_reading;
use crate::{neon_utils, ImageView, ImageViewMut};
#[target_feature(enable = "neon")]
pub(crate) unsafe fn multiply_alpha(
src_view: &impl ImageView<Pixel = U8x2>,
dst_view: &mut impl ImageViewMut<Pixel = U8x2>,
) {
let src_rows = src_view.iter_rows(0);
let dst_rows = dst_view.iter_rows_mut(0);
for (src_row, dst_row) in src_rows.zip(dst_rows) {
multiply_alpha_row(src_row, dst_row);
}
}
#[target_feature(enable = "neon")]
pub(crate) unsafe fn multiply_alpha_inplace(image_view: &mut impl ImageViewMut<Pixel = U8x2>) {
for row in image_view.iter_rows_mut(0) {
multiply_alpha_row_inplace(row);
}
}
#[inline(always)]
unsafe fn multiply_alpha_row(src_row: &[U8x2], dst_row: &mut [U8x2]) {
let src_chunks = src_row.chunks_exact(32);
let src_remainder = src_chunks.remainder();
let mut dst_chunks = dst_row.chunks_exact_mut(32);
let src_dst = src_chunks.zip(&mut dst_chunks);
foreach_with_pre_reading(
src_dst,
|(src, dst)| {
let pixels = neon_utils::load_deintrel_u8x16x4(src, 0);
let dst_ptr = dst.as_mut_ptr() as *mut u8;
(pixels, dst_ptr)
},
|(mut pixels, dst_ptr)| {
pixels = multiplies_alpha_32_pixels(pixels);
vst4q_u8(dst_ptr, pixels);
},
);
let src_chunks = src_remainder.chunks_exact(16);
let src_remainder = src_chunks.remainder();
let dst_reminder = dst_chunks.into_remainder();
let mut dst_chunks = dst_reminder.chunks_exact_mut(16);
let mut src_dst = src_chunks.zip(&mut dst_chunks);
if let Some((src, dst)) = src_dst.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x4(src, 0);
pixels = multiplies_alpha_16_pixels(pixels);
let dst_ptr = dst.as_mut_ptr() as *mut u8;
vst4_u8(dst_ptr, pixels);
}
let src_chunks = src_remainder.chunks_exact(8);
let src_remainder = src_chunks.remainder();
let dst_reminder = dst_chunks.into_remainder();
let mut dst_chunks = dst_reminder.chunks_exact_mut(8);
let mut src_dst = src_chunks.zip(&mut dst_chunks);
if let Some((src, dst)) = src_dst.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x2(src, 0);
pixels = multiplies_alpha_8_pixels(pixels);
let dst_ptr = dst.as_mut_ptr() as *mut u8;
vst2_u8(dst_ptr, pixels);
}
if !src_remainder.is_empty() {
let dst_reminder = dst_chunks.into_remainder();
native::multiply_alpha_row(src_remainder, dst_reminder);
}
}
#[inline(always)]
unsafe fn multiply_alpha_row_inplace(row: &mut [U8x2]) {
let mut chunks = row.chunks_exact_mut(32);
foreach_with_pre_reading(
&mut chunks,
|chunk| {
let pixels = neon_utils::load_deintrel_u8x16x4(chunk, 0);
let dst_ptr = chunk.as_mut_ptr() as *mut u8;
(pixels, dst_ptr)
},
|(mut pixels, dst_ptr)| {
pixels = multiplies_alpha_32_pixels(pixels);
vst4q_u8(dst_ptr, pixels);
},
);
let reminder = chunks.into_remainder();
let mut chunks = reminder.chunks_exact_mut(16);
if let Some(chunk) = chunks.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x4(chunk, 0);
pixels = multiplies_alpha_16_pixels(pixels);
let chunk_ptr = chunk.as_mut_ptr() as *mut u8;
vst4_u8(chunk_ptr, pixels);
}
let reminder = chunks.into_remainder();
let mut chunks = reminder.chunks_exact_mut(8);
if let Some(chunk) = chunks.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x2(chunk, 0);
pixels = multiplies_alpha_8_pixels(pixels);
let chunk_ptr = chunk.as_mut_ptr() as *mut u8;
vst2_u8(chunk_ptr, pixels);
}
let reminder = chunks.into_remainder();
if !reminder.is_empty() {
native::multiply_alpha_row_inplace(reminder);
}
}
#[inline(always)]
unsafe fn multiplies_alpha_32_pixels(mut pixels: uint8x16x4_t) -> uint8x16x4_t {
let zero_u8x16 = vdupq_n_u8(0);
let alpha_u16 = uint16x8x2_t(
vreinterpretq_u16_u8(vzip1q_u8(pixels.1, zero_u8x16)),
vreinterpretq_u16_u8(vzip2q_u8(pixels.1, zero_u8x16)),
);
pixels.0 = neon_utils::mul_color_to_alpha_u8x16(pixels.0, alpha_u16, zero_u8x16);
let alpha_u16 = uint16x8x2_t(
vreinterpretq_u16_u8(vzip1q_u8(pixels.3, zero_u8x16)),
vreinterpretq_u16_u8(vzip2q_u8(pixels.3, zero_u8x16)),
);
pixels.2 = neon_utils::mul_color_to_alpha_u8x16(pixels.2, alpha_u16, zero_u8x16);
pixels
}
#[inline(always)]
unsafe fn multiplies_alpha_16_pixels(mut pixels: uint8x8x4_t) -> uint8x8x4_t {
let zero_u8x8 = vdup_n_u8(0);
let alpha_u16_lo = vreinterpret_u16_u8(vzip1_u8(pixels.1, zero_u8x8));
let alpha_u16_hi = vreinterpret_u16_u8(vzip2_u8(pixels.1, zero_u8x8));
let alpha_u16 = vcombine_u16(alpha_u16_lo, alpha_u16_hi);
pixels.0 = neon_utils::mul_color_to_alpha_u8x8(pixels.0, alpha_u16, zero_u8x8);
let alpha_u16_lo = vreinterpret_u16_u8(vzip1_u8(pixels.3, zero_u8x8));
let alpha_u16_hi = vreinterpret_u16_u8(vzip2_u8(pixels.3, zero_u8x8));
let alpha_u16 = vcombine_u16(alpha_u16_lo, alpha_u16_hi);
pixels.2 = neon_utils::mul_color_to_alpha_u8x8(pixels.2, alpha_u16, zero_u8x8);
pixels
}
#[inline(always)]
unsafe fn multiplies_alpha_8_pixels(mut pixels: uint8x8x2_t) -> uint8x8x2_t {
let zero_u8x8 = vdup_n_u8(0);
let alpha_u16_lo = vreinterpret_u16_u8(vzip1_u8(pixels.1, zero_u8x8));
let alpha_u16_hi = vreinterpret_u16_u8(vzip2_u8(pixels.1, zero_u8x8));
let alpha_u16 = vcombine_u16(alpha_u16_lo, alpha_u16_hi);
pixels.0 = neon_utils::mul_color_to_alpha_u8x8(pixels.0, alpha_u16, zero_u8x8);
pixels
}
#[target_feature(enable = "neon")]
pub(crate) unsafe fn divide_alpha(
src_view: &impl ImageView<Pixel = U8x2>,
dst_view: &mut impl ImageViewMut<Pixel = U8x2>,
) {
let src_rows = src_view.iter_rows(0);
let dst_rows = dst_view.iter_rows_mut(0);
for (src_row, dst_row) in src_rows.zip(dst_rows) {
divide_alpha_row(src_row, dst_row);
}
}
#[target_feature(enable = "neon")]
pub(crate) unsafe fn divide_alpha_inplace(image_view: &mut impl ImageViewMut<Pixel = U8x2>) {
for row in image_view.iter_rows_mut(0) {
divide_alpha_row_inplace(row);
}
}
#[inline(always)]
unsafe fn divide_alpha_row(src_row: &[U8x2], dst_row: &mut [U8x2]) {
let src_chunks = src_row.chunks_exact(16);
let src_remainder = src_chunks.remainder();
let mut dst_chunks = dst_row.chunks_exact_mut(16);
let src_dst = src_chunks.zip(&mut dst_chunks);
foreach_with_pre_reading(
src_dst,
|(src, dst)| {
let pixels = neon_utils::load_deintrel_u8x16x2(src, 0);
let dst_ptr = dst.as_mut_ptr() as *mut u8;
(pixels, dst_ptr)
},
|(mut pixels, dst_ptr)| {
pixels = divide_alpha_16_pixels(pixels);
vst2q_u8(dst_ptr, pixels);
},
);
let src_chunks = src_remainder.chunks_exact(8);
let src_remainder = src_chunks.remainder();
let dst_reminder = dst_chunks.into_remainder();
let mut dst_chunks = dst_reminder.chunks_exact_mut(8);
let mut src_dst = src_chunks.zip(&mut dst_chunks);
if let Some((src, dst)) = src_dst.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x2(src, 0);
pixels = divide_alpha_8_pixels(pixels);
let dst_ptr = dst.as_mut_ptr() as *mut u8;
vst2_u8(dst_ptr, pixels);
}
if !src_remainder.is_empty() {
let dst_reminder = dst_chunks.into_remainder();
let mut src_pixels = [U8x2::new([0; 2]); 8];
src_pixels
.iter_mut()
.zip(src_remainder)
.for_each(|(d, s)| *d = *s);
let mut dst_pixels = [U8x2::new([0; 2]); 8];
let mut pixels = neon_utils::load_deintrel_u8x8x2(&src_pixels, 0);
pixels = divide_alpha_8_pixels(pixels);
let dst_ptr = dst_pixels.as_mut_ptr() as *mut u8;
vst2_u8(dst_ptr, pixels);
dst_pixels
.iter()
.zip(dst_reminder)
.for_each(|(s, d)| *d = *s);
}
}
#[inline(always)]
unsafe fn divide_alpha_row_inplace(row: &mut [U8x2]) {
let mut chunks = row.chunks_exact_mut(16);
foreach_with_pre_reading(
&mut chunks,
|chunk| {
let pixels = neon_utils::load_deintrel_u8x16x2(chunk, 0);
let dst_ptr = chunk.as_mut_ptr() as *mut u8;
(pixels, dst_ptr)
},
|(mut pixels, dst_ptr)| {
pixels = divide_alpha_16_pixels(pixels);
vst2q_u8(dst_ptr, pixels);
},
);
let reminder = chunks.into_remainder();
let mut chunks = reminder.chunks_exact_mut(8);
if let Some(chunk) = chunks.next() {
let mut pixels = neon_utils::load_deintrel_u8x8x2(chunk, 0);
pixels = divide_alpha_8_pixels(pixels);
let chunk_ptr = chunk.as_mut_ptr() as *mut u8;
vst2_u8(chunk_ptr, pixels);
}
let reminder = chunks.into_remainder();
if !reminder.is_empty() {
let mut src_pixels = [U8x2::new([0; 2]); 8];
src_pixels
.iter_mut()
.zip(reminder.iter())
.for_each(|(d, s)| *d = *s);
let mut dst_pixels = [U8x2::new([0; 2]); 8];
let mut pixels = neon_utils::load_deintrel_u8x8x2(&src_pixels, 0);
pixels = divide_alpha_8_pixels(pixels);
let dst_ptr = dst_pixels.as_mut_ptr() as *mut u8;
vst2_u8(dst_ptr, pixels);
dst_pixels.iter().zip(reminder).for_each(|(s, d)| *d = *s);
}
}
#[inline(always)]
unsafe fn divide_alpha_16_pixels(mut pixels: uint8x16x2_t) -> uint8x16x2_t {
let zero = vdupq_n_u8(0);
let alpha_scale = vdupq_n_f32(255.0 * 256.0);
let nonzero_alpha_mask = vmvnq_u8(vceqzq_u8(pixels.1));
let alpha_u16_lo = vzip1q_u8(pixels.1, zero);
let alpha_u16_hi = vzip2q_u8(pixels.1, zero);
let alpha_f32_0 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip1q_u8(alpha_u16_lo, zero)));
let recip_alpha_f32_0 = vdivq_f32(alpha_scale, alpha_f32_0);
let recip_alpha_u16_0 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_0));
let alpha_f32_1 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip2q_u8(alpha_u16_lo, zero)));
let recip_alpha_f32_1 = vdivq_f32(alpha_scale, alpha_f32_1);
let recip_alpha_u16_1 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_1));
let alpha_f32_2 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip1q_u8(alpha_u16_hi, zero)));
let recip_alpha_f32_2 = vdivq_f32(alpha_scale, alpha_f32_2);
let recip_alpha_u16_2 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_2));
let alpha_f32_3 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip2q_u8(alpha_u16_hi, zero)));
let recip_alpha_f32_3 = vdivq_f32(alpha_scale, alpha_f32_3);
let recip_alpha_u16_3 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_3));
let recip_alpha = uint16x8x2_t(
vcombine_u16(recip_alpha_u16_0, recip_alpha_u16_1),
vcombine_u16(recip_alpha_u16_2, recip_alpha_u16_3),
);
pixels.0 = neon_utils::mul_color_recip_alpha_u8x16(pixels.0, recip_alpha, zero);
pixels.0 = vandq_u8(pixels.0, nonzero_alpha_mask);
pixels
}
#[inline(always)]
unsafe fn divide_alpha_8_pixels(mut pixels: uint8x8x2_t) -> uint8x8x2_t {
let zero_u8x8 = vdup_n_u8(0);
let zero_u8x16 = vdupq_n_u8(0);
let alpha_scale = vdupq_n_f32(255.0 * 256.0);
let nonzero_alpha_mask = vmvn_u8(vceqz_u8(pixels.1));
let alpha_u16_lo = vzip1_u8(pixels.1, zero_u8x8);
let alpha_u16_hi = vzip2_u8(pixels.1, zero_u8x8);
let alpha_u16 = vcombine_u8(alpha_u16_lo, alpha_u16_hi);
let alpha_f32_0 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip1q_u8(alpha_u16, zero_u8x16)));
let recip_alpha_f32_0 = vdivq_f32(alpha_scale, alpha_f32_0);
let recip_alpha_u16_0 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_0));
let alpha_f32_1 = vcvtq_f32_u32(vreinterpretq_u32_u8(vzip2q_u8(alpha_u16, zero_u8x16)));
let recip_alpha_f32_1 = vdivq_f32(alpha_scale, alpha_f32_1);
let recip_alpha_u16_1 = vmovn_u32(vcvtaq_u32_f32(recip_alpha_f32_1));
let recip_alpha = vcombine_u16(recip_alpha_u16_0, recip_alpha_u16_1);
pixels.0 = neon_utils::mul_color_recip_alpha_u8x8(pixels.0, recip_alpha, zero_u8x8);
pixels.0 = vand_u8(pixels.0, nonzero_alpha_mask);
pixels
}