Llama Runner
A straightforward Rust library for running llama.cpp models locally on device.
Example
// Download and run Gemma3 1B QAT
let runner = default.await.unwrap;
let answer = runner
.get_lm_response
.unwrap;
assert!;
// Download and run Qwen 3.5 4B QAT
// Note that `Gemma3VisionRunner` merely means it's capable
// of running Gemma3 vision models, not necessarily Gemma though
// Configurable using the ::new constructor
let runner = default.await.unwrap;
let eiffel_tower_im =
load_from_memory.unwrap;
let answer = runner
.get_vlm_response
.unwrap;
assert!;
Credits
- llama-cpp-rs: this library is bascially a higher level wrapper around it