Expand description
INT8 Quantized Kernels Module
Provides SIMD-optimized INT8 kernels for:
- 2D Convolution (standard and depthwise)
- Matrix multiplication (GEMM)
- Dot product operations
§Architecture Support
- x86_64: AVX2 with
_mm256_maddubs_epi16and_mm256_madd_epi16 - aarch64: ARM NEON with
vmull_s8andvpadalq_s16 - wasm32: WebAssembly SIMD128 with
i8x16operations