baochuan 0.1.0

A multi-provider AI API client for Rust — connecting your code to every major AI provider.
Documentation

baochuan · 宝船

A multi-provider AI API client for Rust.

Crates.io docs.rs License: MIT


The Name

"宝船" (Bǎochuán) — Treasure Ship

In the early 15th century, Admiral Zheng He commanded the largest wooden fleet the world had ever seen. His colossal treasure ships — some reportedly over 400 feet long — sailed from China to the Persian Gulf, East Africa, and beyond, connecting civilizations that had never met. They carried silk, porcelain, and ideas across the Indian Ocean, opening the world to each other.

baochuan aspires to the same role: a vessel that carries your Rust code to every major AI provider through a single unified interface, regardless of where those providers sail.

The sister library ZhengHe does the same for Java — named after the explorer who famously introduced China to Java (the island). ZhengHe connects Java to the DeepSeek API; baochuan carries the voyage forward in Rust, with a whole fleet of providers.


Features

  • Fully async — built on tokio and reqwest
  • Multi-provider — swap providers without changing your business logic
  • Streaming — native SSE streaming support for real-time token delivery
  • Builder pattern — ergonomic, validated request construction
  • Own implementation — no third-party SDK wrappers; direct HTTP to each provider
  • Extensible — implement the Provider trait to add any provider

Supported Providers

Provider Chat Streaming Model List API Env Var
OpenAI OpenAI native OPENAI_API_KEY
Anthropic Anthropic native ANTHROPIC_API_KEY
Google Gemini Gemini native GEMINI_API_KEY
xAI Grok xAI native XAI_API_KEY
Mistral Mistral native MISTRAL_API_KEY
DeepSeek DeepSeek native DEEPSEEK_API_KEY
GitHub Copilot GitHub Copilot native GITHUB_TOKEN
OpenRouter OpenRouter native OPENROUTER_API_KEY
Moonshot AI / Kimi Moonshot native MOONSHOT_API_KEY
Perplexity Perplexity native PERPLEXITY_API_KEY
Alibaba Qwen DashScope native DASHSCOPE_API_KEY
Cloudflare Workers AI CF native /ai/run/ CLOUDFLARE_ACCOUNT_ID + CLOUDFLARE_API_TOKEN
LM Studio LM Studio /api/v0/ (none)
Ollama Ollama /api/ (none)
llama.cpp llama-server /v1/ (none)
More coming…

Installation

Add baochuan to your Cargo.toml:

[dependencies]
baochuan = "0.1"
tokio = { version = "1", features = ["full"] }

Quickstart

DeepSeek

use baochuan::{providers::DeepSeekProvider, ChatMessage, ChatRequestBuilder, Provider};

#[tokio::main]
async fn main() {
    let provider = DeepSeekProvider::new(
        std::env::var("DEEPSEEK_API_KEY").expect("DEEPSEEK_API_KEY not set"),
    );

    let request = ChatRequestBuilder::new("deepseek-chat")
        .message(ChatMessage::user("Tell me about the treasure ships of Zheng He."))
        .max_tokens(512)
        .build()
        .unwrap();

    let response = provider.chat(&request).await.unwrap();
    println!("{}", response.content().unwrap_or("(no response)"));
}

OpenRouter

use baochuan::{providers::OpenRouterProvider, ChatMessage, ChatRequestBuilder, Provider};

#[tokio::main]
async fn main() {
    let provider = OpenRouterProvider::new(
        std::env::var("OPENROUTER_API_KEY").expect("OPENROUTER_API_KEY not set"),
    )
    .site_name("My App");

    // OpenRouter gives you access to hundreds of models:
    let request = ChatRequestBuilder::new("anthropic/claude-3-5-sonnet")
        .message(ChatMessage::user("What is the speed of light?"))
        .build()
        .unwrap();

    let response = provider.chat(&request).await.unwrap();
    println!("{}", response.content().unwrap_or("(no response)"));
}

Streaming

use baochuan::{providers::DeepSeekProvider, ChatMessage, ChatRequestBuilder, Provider};
use futures_util::StreamExt;

#[tokio::main]
async fn main() {
    let provider = DeepSeekProvider::new(
        std::env::var("DEEPSEEK_API_KEY").unwrap(),
    );

    let request = ChatRequestBuilder::new("deepseek-chat")
        .message(ChatMessage::user("Write a haiku about Rust."))
        .build()
        .unwrap();

    let mut stream = provider.stream_chat(&request).await.unwrap();

    while let Some(chunk) = stream.next().await {
        let chunk = chunk.unwrap();
        if let Some(text) = chunk.delta_content() {
            print!("{text}");
        }
    }
    println!();
}

Provider-agnostic code

use baochuan::{ChatMessage, ChatRequestBuilder, Provider};

async fn ask(provider: &dyn Provider, model: &str, question: &str) -> String {
    let request = ChatRequestBuilder::new(model)
        .message(ChatMessage::user(question))
        .build()
        .unwrap();

    provider
        .chat(&request)
        .await
        .unwrap()
        .content()
        .unwrap_or("")
        .to_string()
}

Conversation with System Prompt

use baochuan::{providers::DeepSeekProvider, ChatMessage, ChatRequestBuilder, Provider};

#[tokio::main]
async fn main() {
    let provider = DeepSeekProvider::new(std::env::var("DEEPSEEK_API_KEY").unwrap());

    let request = ChatRequestBuilder::new("deepseek-chat")
        .message(ChatMessage::system("You are a concise assistant. Reply in one sentence."))
        .message(ChatMessage::user("What is baochuan?"))
        .temperature(0.7)
        .max_tokens(128)
        .build()
        .unwrap();

    let response = provider.chat(&request).await.unwrap();
    println!("{}", response.content().unwrap_or(""));
}

Adding a Provider

Implement the Provider trait for any HTTP-based AI API:

use async_trait::async_trait;
use baochuan::{
    BaochuanError, ChatRequest, ChatResponse, Provider,
    provider::ChunkStream,
};

pub struct MyProvider { /* ... */ }

#[async_trait]
impl Provider for MyProvider {
    fn name(&self) -> &str { "my-provider" }

    async fn chat(&self, request: &ChatRequest) -> Result<ChatResponse, BaochuanError> {
        todo!()
    }

    async fn stream_chat(&self, request: &ChatRequest) -> Result<ChunkStream, BaochuanError> {
        todo!()
    }
}

Security

  • Never hard‑code API keys. Always read them from environment variables or a secrets manager.
  • Use .gitignore to exclude .env files from version control.
  • Rotate keys immediately if you suspect they have been exposed.
export DEEPSEEK_API_KEY="sk-..."
export OPENROUTER_API_KEY="sk-or-..."

License

MIT — see LICENSE.