infernum-server 0.2.0-rc.2

HTTP API server for local LLM inference