[](https://crates.io/crates/curtana)
[](https://docs.rs/curtana/)
[](https://github.com/with-caer/curtana/blob/main/LICENSE.txt)
Simplified, low-overhead wrapper over [`llama.cpp`](https://github.com/ggml-org/llama.cpp)
powered by the [`llama-cpp-2`](https://github.com/utilityai/llama-cpp-rs/tree/main) crate
supporting most `.gguf` formatted "Chat" and "Embedding" models.
## Build and Test
0. Install `cmake` and [`rust`](https://www.rust-lang.org/tools/install) on your system.
1. Download the GGUF models used during testing:
- `wget https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q6_K.gguf`
- `wget https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF/resolve/main/nomic-embed-text-v1.5.f16.gguf`
2. Run `cargo test`
4. ???
5. Profit.
## License
Copyright © 2025 With Caer, LLC.
Licensed under the MIT license. Refer to [the license file](https://github.com/with-caer/curtana/blob/main/LICENSE.txt) for more info.