hoosh 1.3.0

AI inference gateway — multi-provider LLM routing, local model serving, speech-to-text, and token budget management
Documentation
//! llama.cpp provider — thin wrapper over OpenAI-compatible API.

use crate::inference::{InferenceRequest, InferenceResponse, ModelInfo};
use crate::provider::openai_compat::OpenAiCompatibleProvider;
use crate::provider::{LlmProvider, ProviderType, TlsConfig};

/// llama.cpp server provider (default: `http://localhost:8080`).
pub struct LlamaCppProvider {
    inner: OpenAiCompatibleProvider,
}

impl LlamaCppProvider {
    pub fn new(base_url: impl Into<String>, tls_config: Option<&TlsConfig>) -> Self {
        let url = base_url.into();
        let url = if url.is_empty() {
            "http://localhost:8080".to_string()
        } else {
            url
        };
        Self {
            inner: OpenAiCompatibleProvider::new(url, None, ProviderType::LlamaCpp, tls_config),
        }
    }
}

#[async_trait::async_trait]
impl LlmProvider for LlamaCppProvider {
    async fn infer(&self, request: &InferenceRequest) -> anyhow::Result<InferenceResponse> {
        self.inner.infer(request).await
    }

    async fn infer_stream(
        &self,
        request: InferenceRequest,
    ) -> anyhow::Result<tokio::sync::mpsc::Receiver<anyhow::Result<String>>> {
        self.inner.infer_stream(request).await
    }

    async fn list_models(&self) -> anyhow::Result<Vec<ModelInfo>> {
        self.inner.list_models().await
    }

    async fn health_check(&self) -> anyhow::Result<bool> {
        self.inner.health_check().await
    }

    fn provider_type(&self) -> ProviderType {
        ProviderType::LlamaCpp
    }
}

#[cfg(test)]
mod tests {
    use super::*;

    #[test]
    fn default_url() {
        let p = LlamaCppProvider::new("", None);
        assert_eq!(p.inner.base_url(), "http://localhost:8080");
    }

    #[test]
    fn custom_url() {
        let p = LlamaCppProvider::new("http://gpu-box:9999", None);
        assert_eq!(p.inner.base_url(), "http://gpu-box:9999");
    }

    #[test]
    fn provider_type_is_llamacpp() {
        let p = LlamaCppProvider::new("", None);
        assert_eq!(p.provider_type(), ProviderType::LlamaCpp);
    }
}