Expand description
SIMD-accelerated operations SIMD-accelerated quantization operations
This module provides optimized SIMD implementations for performance-critical quantization operations using the scirs2-core SIMD abstraction layer.
§Features
- Vectorized Quantization: SIMD-accelerated per-tensor quantization
- Vectorized Dequantization: SIMD-accelerated dequantization operations
- Fast Min/Max Finding: Hardware-accelerated min/max computation for calibration
- Batch Operations: Optimized batch processing for multiple tensors
- Fallback Support: Automatic fallback to scalar operations when SIMD unavailable
Structs§
- Mobile
Optimization Hints - Mobile optimization configuration hints
- Tensor
Stats - Tensor statistics structure
Functions§
- calculate_
tensor_ stats_ simd - SIMD-accelerated statistics calculation for quantization calibration
- dequantize_
per_ tensor_ affine_ simd - SIMD-accelerated per-tensor dequantization
- find_
min_ max_ simd - SIMD-accelerated min/max finding for calibration
- get_
mobile_ optimization_ hints - Get mobile-specific optimization recommendations
- get_
simd_ width - Get optimal SIMD vector width for current hardware
- is_
simd_ available - Check if SIMD operations are available on current hardware
- quantize_
batch_ consistent_ simd - SIMD-accelerated batch quantization for consistent parameters
- quantize_
mobile_ optimized - Mobile-optimized quantization with reduced memory usage
- quantize_
per_ channel_ simd - SIMD-accelerated per-channel quantization
- quantize_
per_ tensor_ affine_ simd - SIMD-accelerated per-tensor quantization
- quantize_
to_ int8_ simd - SIMD-accelerated floating-point to integer quantization (optimized for INT8)