omnillm 0.1.5

Production-grade LLM API gateway with multi-key load balancing, per-key rate limiting, circuit breaking, and cost tracking
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
mod cache;
mod chat;
mod responses;
mod tools;

pub(in crate::protocol::openai) use cache::{
    emit_openai_prompt_cache_policy, openai_chat_usage_json, openai_responses_usage_json,
    parse_openai_prompt_cache_policy, parse_openai_prompt_cache_usage,
};
pub(in crate::protocol::openai) use chat::{
    openai_chat_message_json, parse_openai_chat_message, parse_openai_chat_structured_output,
};
pub(in crate::protocol::openai) use responses::{
    emit_openai_responses_capabilities, openai_responses_input_item, openai_responses_output_item,
    parse_openai_responses_capabilities, parse_openai_responses_input,
    parse_openai_responses_output, parse_openai_responses_single_output_item,
};
pub(in crate::protocol::openai) use tools::{emit_function_tools, parse_function_tools};