infernum-server 0.2.0-rc.2

HTTP API server for local LLM inference

infernum-server

There is very little structured metadata to build this page from currently. You should check the main library docs, readme, or Cargo.toml in case the author documented the features in them.

This version has 10 feature flags, 0 of them enabled by default.

default

This feature flag does not enable additional features.

accelerate

cuda

cuda-full

flash-attn

haagenti-gpu

holotensor

integration-tests

This feature flag does not enable additional features.

lance

metal

mkl