[package]
edition = "2024"
name = "hot-loop"
version = "0.1.6"
build = false
autolib = false
autobins = false
autoexamples = false
autotests = false
autobenches = false
description = "Running Gguf Chat-Models on Pure-Rust, Uses the Candle Backend"
readme = "README.md"
keywords = [
"llm",
"gguf",
"candle",
"inference",
"model",
]
license = "MIT"
repository = "https://github.com/Qu1ntev/hot-loop"
[features]
accelerate = [
"dep:accelerate-src",
"candle-core/accelerate",
"candle-transformers/accelerate",
"candle-nn/accelerate",
]
cuda = [
"candle-core/cuda",
"candle-transformers/cuda",
"candle-nn/cuda",
]
default = []
mkl = [
"dep:intel-mkl-src",
"candle-core/mkl",
"candle-transformers/mkl",
"candle-nn/mkl",
]
[lib]
name = "hot_loop"
path = "src/lib.rs"
[dependencies.accelerate-src]
version = "0.3.2"
optional = true
[dependencies.candle-core]
version = "0.9.2"
[dependencies.candle-nn]
version = "0.9.2"
[dependencies.candle-transformers]
version = "0.9.2"
[dependencies.intel-mkl-src]
version = "0.8.1"
optional = true
[dependencies.thiserror]
version = "2.0.18"
[dependencies.tokenizers]
version = "0.22.2"