Crate llm_base

source ·
Expand description

LLaMA-rs is a Rust port of the llama.cpp project. This allows running inference for Facebook’s LLaMA model on a CPU with good performance using full precision, f16 or 4-bit quantized versions of the model.

Re-exports

Modules

  • Large language model traits and types
  • Utilities for interacting with LLMs and loading them.

Structs

Enums

Traits

Functions

  • Load a GGML model from the path and configure it per the params. The status of the loading process will be reported through load_progress_callback.
  • A implementation for load_progress_callback that outputs to stdout.
  • Quantizes a model.

Type Definitions

  • The identifier of a token in a vocabulary.