Skip to main content

Module simd_ops

Module simd_ops 

Source
Expand description

SIMD-accelerated operations SIMD-accelerated quantization operations

This module provides optimized SIMD implementations for performance-critical quantization operations using the scirs2-core SIMD abstraction layer.

§Features

  • Vectorized Quantization: SIMD-accelerated per-tensor quantization
  • Vectorized Dequantization: SIMD-accelerated dequantization operations
  • Fast Min/Max Finding: Hardware-accelerated min/max computation for calibration
  • Batch Operations: Optimized batch processing for multiple tensors
  • Fallback Support: Automatic fallback to scalar operations when SIMD unavailable

Structs§

MobileOptimizationHints
Mobile optimization configuration hints
TensorStats
Tensor statistics structure

Functions§

calculate_tensor_stats_simd
SIMD-accelerated statistics calculation for quantization calibration
dequantize_per_tensor_affine_simd
SIMD-accelerated per-tensor dequantization
find_min_max_simd
SIMD-accelerated min/max finding for calibration
get_mobile_optimization_hints
Get mobile-specific optimization recommendations
get_simd_width
Get optimal SIMD vector width for current hardware
is_simd_available
Check if SIMD operations are available on current hardware
quantize_batch_consistent_simd
SIMD-accelerated batch quantization for consistent parameters
quantize_mobile_optimized
Mobile-optimized quantization with reduced memory usage
quantize_per_channel_simd
SIMD-accelerated per-channel quantization
quantize_per_tensor_affine_simd
SIMD-accelerated per-tensor quantization
quantize_to_int8_simd
SIMD-accelerated floating-point to integer quantization (optimized for INT8)