openai-interface
A low-level Rust interface for interacting with OpenAI's API. Both streaming and non-streaming APIs are supported.
Currently only chat completion is supported. FIM completion, image generation, etc. are still in development.
Features
- Chat Completions: Full support for OpenAI's chat completion API
- Streaming and Non-streaming: Support for both streaming and non-streaming responses
- Strong Typing: Complete type definitions for all API requests and responses, utilizing Rust's powerful typing system
- Error Handling: All errors are converted into
crate::error::RequestError
orcrate::error::ResponseError
- Async/Await: Built with async/await support
- Musl Support: Designed to work with musl libc out-of-the-box
- Multiple Provider Support: Works with OpenAI, DeepSeek, Qwen, and other compatible APIs
Installation
[!WARNING] Versions prior to 0.3.0 has serious issues on processing SSE streaming responses. Instead of a single chunk, chances are that multiple chunks will be returned in each iteration of
chat::request::ChatCompletion::get_streaming_response
.
Add this to your Cargo.toml
:
[]
= "0.3"
Usage
Chat Completion
Non-streaming Chat Completion
use LazyLock;
use ;
use Completion;
// You need to provide your own DeepSeek API key at /keys/deepseek_domestic_key
const DEEPSEEK_API_KEY: =
new;
const DEEPSEEK_CHAT_URL: &'static str = "https://api.deepseek.com/chat/completions";
const DEEPSEEK_MODEL: &'static str = "deepseek-chat";
async
Streaming Chat Completion
use ;
use Completion;
use StreamExt;
use LazyLock;
// You need to provide your own DeepSeek API key at /keys/deepseek_domestic_key
const DEEPSEEK_API_KEY: =
new;
const DEEPSEEK_CHAT_URL: &'static str = "https://api.deepseek.com/chat/completions";
const DEEPSEEK_MODEL: &'static str = "deepseek-chat";
async
Custom Request Parameters
You can customize whatever request parameters you want. If you need extra platform
-specific fields, find them in extra_body
, or add them to extra_body_map
.
Getting Response
Parse the response content by crate::chat::response::streaming::Completion::parse_string
or crate::chat::response::no_streaming::Completion::parse_string
Error Handling
All errors are converted into crate::error::RequestError
or
crate::error::ResponseError
.
Musl Build
This crate is designed to adapt with the musl libc, making it suitable for lightweight deployments in containerized environments. Longer compile times may be required, for openssl is needed to be built from source.
To build for musl:
Supported Providers
This crate aims to support standard OpenAI-compatible API endpoints. Sadly, the People's Republic of China is blocked by the OpenAI website, and I can only refer to Manuals from DeepSeek and Qwen. Please open an issue if there are any mistakes or inaccuracies in my implementation.
Contributing
Contributions are welcome! Please feel free to submit pull requests or open issues for bugs and feature requests.
License
This project is licensed under the AGPL-3.0 License - see the LICENSE file for details.