OpenAI API for Rust

A community-maintained library provides a simple and convenient way to interact with the OpenAI API.
No complex async and redundant dependencies.
API
check official API reference
API |
Support |
Models |
✔️ |
Completions |
✔️ |
Chat |
✔️ |
Edits |
✔️ |
Images |
✔️ |
Embeddings |
✔️ |
Audio |
✔️ |
Files |
❌ |
Fine-tunes |
❌ |
Moderations |
❌ |
Engines |
❌ |
Usage
Add the following to your Cargo.toml file:
openai_api_rust = "0.1.3"
Export your API key into the environment variables
export OPENAI_API_KEY=<your_api_key>
Then use the crate in your Rust code:
use openai_api_rust::*;
use openai_api_rust::chat::*;
use openai_api_rust::completions::*;
fn main() {
let auth = Auth::from_env().unwrap();
let openai = OpenAI::new(auth, "https://api.openai.com/v1/");
let body = ChatBody {
model: "gpt-3.5-turbo".to_string(),
max_tokens: Some(7),
temperature: Some(0_f32),
top_p: Some(0_f32),
n: Some(2),
stream: Some(false),
stop: None,
presence_penalty: None,
frequency_penalty: None,
logit_bias: None,
user: None,
messages: vec![Message {
role: Some("user".to_string()),
content: Some("Hello!".to_string()),
}],
};
let rs = openai.chat_completion_create(&body);
let choice = rs.unwrap().choices;
let message = &choice[0].message.as_ref().unwrap();
let content = message.content.as_ref().unwrap();
assert!(content.contains("Hello"));
}
Use proxy
Load proxy from env
let openai = OpenAI::new(auth, "https://api.openai.com/v1/")
.use_env_proxy()
.unwrap();
Set the proxy manually
let openai = OpenAI::new(auth, "https://api.openai.com/v1/")
.set_proxy("http://127.0.0.1:1080");
License
This library is distributed under the terms of the MIT license. See LICENSE for details.