//! # Knowledge Distillation
//!
//! Compress models by transferring knowledge from a large teacher to a small student.
//!
//! | Module | Contents |
//! |------------|------------------------------------------------------|
//! | `loss` | [`DistilLoss`] — KL, MSE, cosine, combined losses |
//! | `response` | [`ResponseDistiller`] — soft + hard label training |
//! | `feature` | [`FeatureDistiller`] — intermediate activation matching |
pub use FeatureDistiller;
pub use ;
pub use ResponseDistiller;