pub fn quantize_per_tensor( tensor: &Tensor, scale: f32, zero_point: i32, _dtype: DType, ) -> Result<Tensor>
Quantize a tensor to INT8 using specified scale and zero point