kubellm 0.0.1

KubeLLM is a simple LLM proxy designed to run on Kubernetes or standalone
Documentation

KubeLLM

An opinionated LLM proxy written in Rust to be used with Kubernetes or standalone.

[!WARNING] Under development, not ready for production.

Design goals

  • An API that allows calling different LLM providers based on the OpenAI spec
  • Transparent configuration in code
  • Provides a way to generate virtual keys in Kubernetes
  • Spend and usage tracking
  • Logging into local database
  • Prometheus metrics

Personal goals

  • Learning Rust