rig-llama-cpp 0.1.4

Rig completion provider for local GGUF models via llama.cpp, with streaming, tool calling, reasoning, and multimodal (mtmd) support.
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
//! Single integration binary covering all model-bearing tests.
//!
//! All tests in this binary are `#[ignore]` and download GGUF fixtures
//! via `hf-hub` on first run (cached at `~/.cache/huggingface/hub`).
//! Plan for ~20 GB of downloads on a cold cache.
//!
//! Run with:
//!
//! ```sh
//! cargo test --test e2e --features mtmd -- --ignored --nocapture
//! ```

mod common;
mod embedding;
mod gemma;
mod qwen;
mod reload;