use vortex_error::VortexResult;
use crate::aliases::hash_set::HashSet;
use crate::stats::PRUNING_STATS;
use crate::{Array, EncodingId};
pub trait CompressionStrategy {
fn compress(&self, array: &Array) -> VortexResult<Array>;
fn used_encodings(&self) -> HashSet<EncodingId>;
}
pub fn check_validity_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
use vortex_error::VortexExpect;
let old_validity = arr
.logical_validity()
.vortex_expect("failed to compute validity")
.len();
let new_validity = compressed
.logical_validity()
.vortex_expect("failed to compute validity ")
.len();
debug_assert!(
old_validity == new_validity,
"validity length changed after compression: {old_validity} -> {new_validity}\n From tree {} To tree {}\n",
arr.tree_display(),
compressed.tree_display()
);
}
}
pub fn check_dtype_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
debug_assert!(
arr.dtype() == compressed.dtype(),
"Compression changed dtype: {} -> {}\nFrom array: {}Into array {}",
arr.dtype(),
compressed.dtype(),
arr.tree_display(),
compressed.tree_display(),
);
}
}
pub fn check_statistics_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
use vortex_scalar::Scalar;
use crate::stats::Stat;
for (stat, value) in arr
.statistics()
.to_set()
.into_iter()
.filter(|(stat, _)| *stat != Stat::RunCount)
{
let compressed_scalar = compressed
.statistics()
.get(stat)
.map(|sv| Scalar::new(stat.dtype(compressed.dtype()), sv));
debug_assert_eq!(
compressed_scalar,
Some(Scalar::new(stat.dtype(arr.dtype()), value.clone())),
"Compression changed {stat} from {value} to {}",
compressed_scalar
.as_ref()
.map(|s| s.to_string())
.unwrap_or_else(|| "null".to_string()),
);
}
}
}
pub fn compute_precompression_stats(arr: &Array) -> VortexResult<()> {
arr.statistics().compute_uncompressed_size_in_bytes();
arr.statistics().compute_all(PRUNING_STATS).map(|_| ())
}