ambi 0.3.1

A flexible, multi-backend, customizable AI agent framework, entirely based on Rust.
Documentation
// src/llm/providers/llama_cpp/session.rs

use llama_cpp_2::token::LlamaToken;

/// Holds mutable generation state that persists between requests.
///
/// Note: Model, context and batch are owned outside this struct to avoid
/// self‑referential lifetimes.
pub(crate) struct InferenceSession {
    /// Tokens of the last processed prompt plus generated tokens.
    pub history_tokens: Vec<LlamaToken>,
    /// Incomplete UTF‑8 byte sequence from the previous token decode.
    pub utf8_buffer: Vec<u8>,
    /// Current position in the sequence (number of tokens already processed).
    pub pos: i32,
}

impl InferenceSession {
    pub fn new() -> Self {
        Self {
            history_tokens: Vec::new(),
            utf8_buffer: Vec::with_capacity(32),
            pos: 0,
        }
    }

    /// Create a snapshot of the current state
    pub fn snapshot(&self) -> (Vec<LlamaToken>, Vec<u8>, i32) {
        (
            self.history_tokens.clone(),
            self.utf8_buffer.clone(),
            self.pos,
        )
    }

    /// Restore state from snapshot
    pub fn restore(&mut self, snapshot: (Vec<LlamaToken>, Vec<u8>, i32)) {
        self.history_tokens = snapshot.0;
        self.utf8_buffer = snapshot.1;
        self.pos = snapshot.2;
    }

    pub fn reset(&mut self) {
        self.history_tokens.clear();
        self.utf8_buffer.clear();
        self.pos = 0;
    }
}