//! Llama.cpp local inference engine implementation using direct C bindings.
//!
//! This module provides a complete, high-performance local LLM backend backed by
//! the `llama.cpp` library, supporting GPU offloading, KV-cache management,
//! multimodal (vision) inference via MTMD, and entropy evaluation.
pub
pub
pub
pub
pub
pub
pub
pub
pub use LlamaEngine;