use super::{helpers, Inferer};
use crate::{batcher::ScratchPadView, model_api::ModelApi};
use anyhow::Result;
use parking_lot::{RwLock, RwLockReadGuard, RwLockUpgradableReadGuard, RwLockWriteGuard};
use std::{
collections::{hash_map::Entry, HashMap},
ops::Deref,
};
use tract_core::prelude::*;
use tract_hir::prelude::*;
pub struct MemoizingDynamicInferer {
symbol: Symbol,
model: TypedModel,
model_api: ModelApi,
model_cache: RwLock<HashMap<usize, TypedSimplePlan<TypedModel>>>,
}
impl MemoizingDynamicInferer {
pub fn from_model(model: InferenceModel, preloaded_sizes: &[usize]) -> TractResult<Self> {
let model_api = ModelApi::for_model(&model)?;
let (symbol, model) = helpers::build_symbolic_model(model, &model_api.inputs)?;
let this = Self {
symbol,
model,
model_api,
model_cache: Default::default(),
};
for size in preloaded_sizes {
this.get_concrete_model(*size)?;
}
Ok(this)
}
pub fn from_typed(mut model: TypedModel, preloaded_sizes: &[usize]) -> TractResult<Self> {
let model_api = ModelApi::for_typed_model(&model)?;
let symbol = helpers::build_symbolic_typed(&mut model)?;
let this = Self {
symbol,
model,
model_api,
model_cache: Default::default(),
};
for size in preloaded_sizes {
this.get_concrete_model(*size)?;
}
Ok(this)
}
fn build_inputs(&self, batch: &mut ScratchPadView<'_>) -> Result<TVec<TValue>> {
let size = batch.len();
let mut inputs = TVec::default();
for (idx, (name, shape)) in self.model_api.inputs.iter().enumerate() {
assert_eq!(name, batch.input_name(idx));
let mut full_shape = tvec![size];
full_shape.extend_from_slice(shape);
let total_count: usize = full_shape.iter().product();
assert_eq!(total_count, batch.input_slot(idx).len());
let shape = full_shape;
let tensor = Tensor::from_shape(&shape, batch.input_slot(idx))?;
inputs.push(tensor.into());
}
Ok(inputs)
}
fn get_concrete_model(
&self,
size: usize,
) -> Result<impl Deref<Target = TypedSimplePlan<TypedModel>> + '_> {
let cache = self.model_cache.upgradable_read();
let cache = {
if !cache.contains_key(&size) {
let mut content = RwLockUpgradableReadGuard::upgrade(cache);
if let Entry::Vacant(e) = content.entry(size) {
let p = self
.model
.concretize_dims(&SymbolValues::default().with(&self.symbol, size as i64))?
.into_optimized()?
.into_decluttered()?
.into_runnable()?;
e.insert(p);
}
RwLockWriteGuard::downgrade(content)
} else {
RwLockUpgradableReadGuard::downgrade(cache)
}
};
Ok(RwLockReadGuard::map(cache, |c| &c[&size]))
}
}
impl Inferer for MemoizingDynamicInferer {
fn select_batch_size(&self, max_count: usize) -> usize {
max_count
}
fn infer_raw(&self, pad: &mut ScratchPadView<'_>) -> Result<(), anyhow::Error> {
let count = pad.len();
let inputs = self.build_inputs(pad)?;
let result = self.get_concrete_model(count)?.run(inputs)?;
for idx in 0..self.model_api.outputs.len() {
let value = result[idx].as_slice::<f32>()?;
pad.output_slot_mut(idx).copy_from_slice(value);
}
Ok(())
}
fn raw_input_shapes(&self) -> &[(String, Vec<usize>)] {
&self.model_api.inputs
}
fn raw_output_shapes(&self) -> &[(String, Vec<usize>)] {
&self.model_api.outputs
}
fn begin_agent(&self, _id: u64) {}
fn end_agent(&self, _id: u64) {}
}