use vortex_error::VortexResult;
use crate::aliases::hash_set::HashSet;
use crate::stats::PRUNING_STATS;
use crate::{Array, EncodingId};
pub trait CompressionStrategy {
fn compress(&self, array: &Array) -> VortexResult<Array>;
fn used_encodings(&self) -> HashSet<EncodingId>;
}
pub fn check_validity_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
use vortex_error::VortexExpect;
let old_validity = arr
.validity_mask()
.vortex_expect("failed to compute validity")
.len();
let new_validity = compressed
.validity_mask()
.vortex_expect("failed to compute validity ")
.len();
debug_assert!(
old_validity == new_validity,
"validity length changed after compression: {old_validity} -> {new_validity}\n From tree {} To tree {}\n",
arr.tree_display(),
compressed.tree_display()
);
}
}
pub fn check_dtype_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
debug_assert!(
arr.dtype() == compressed.dtype(),
"Compression changed dtype: {} -> {}\nFrom array: {}Into array {}",
arr.dtype(),
compressed.dtype(),
arr.tree_display(),
compressed.tree_display(),
);
}
}
pub fn check_statistics_unchanged(arr: &Array, compressed: &Array) {
let _ = arr;
let _ = compressed;
#[cfg(debug_assertions)]
{
use crate::stats::Stat;
for (stat, value) in arr
.statistics()
.to_set()
.into_iter()
.filter(|(stat, _)| *stat != Stat::RunCount)
{
let compressed_scalar = compressed
.statistics()
.get(stat)
.map(|sv| sv.into_scalar(stat.dtype(compressed.dtype())));
debug_assert_eq!(
compressed_scalar.clone(),
Some(value.clone().into_scalar(stat.dtype(arr.dtype()))),
"Compression changed {stat} from {value} to {:?}",
compressed_scalar.as_ref(),
);
}
}
}
pub fn compute_precompression_stats(arr: &Array) -> VortexResult<()> {
arr.statistics().compute_uncompressed_size_in_bytes();
arr.statistics().compute_all(PRUNING_STATS).map(|_| ())
}