Expand description
Shared model download and cache layer for Blazen local-inference backends.
Provides ModelCache for downloading and caching ML models from
HuggingFace Hub. Designed to be shared by all
local-inference backends (fastembed, mistral.rs, whisper.cpp, etc.).
Structs§
- Model
Cache - Local cache for ML models downloaded from
HuggingFaceHub.
Enums§
- Cache
Error - Errors that can occur during model cache operations.
Traits§
- Progress
Callback - Callback trait for receiving download progress updates.