Please check the build logs for more information.
See Builds for ideas on how to fix a failed build, or Metadata for how to configure docs.rs builds.
If you believe this is docs.rs' fault, open an issue.
adk-model
LLM model integrations for Rust Agent Development Kit (ADK-Rust) with Gemini, OpenAI, Anthropic, and DeepSeek.
Overview
adk-model provides LLM integrations for the Rust Agent Development Kit (ADK-Rust). Supports all major providers:
- Gemini - Google's Gemini models (3 Pro, 3 Flash, 2.5 Pro, 2.5 Flash, etc.)
- OpenAI - GPT-5.1, GPT-5, GPT-5 Mini, GPT-4o (legacy)
- Anthropic - Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Claude 4
- DeepSeek - DeepSeek R1, DeepSeek V3.1, DeepSeek-Chat with thinking mode
- Groq - Ultra-fast inference (LLaMA 3.3, Mixtral, Gemma)
- Ollama - Local LLMs (LLaMA, Mistral, Qwen, Gemma, etc.)
- Streaming - Real-time response streaming for all providers
- Multimodal - Text, images, audio, video, and PDF input
The crate implements the Llm trait from adk-core, allowing models to be used interchangeably.
Installation
[]
= "0.3.1"
Or use the meta-crate:
[]
= { = "0.3.1", = ["models"] }
Quick Start
Gemini (Google)
use GeminiModel;
use LlmAgentBuilder;
use Arc;
async
OpenAI
use ;
use LlmAgentBuilder;
use Arc;
async
Anthropic (Claude)
use ;
use LlmAgentBuilder;
use Arc;
async
DeepSeek
use ;
use LlmAgentBuilder;
use Arc;
async
Groq (Ultra-Fast)
use ;
use LlmAgentBuilder;
use Arc;
async
Ollama (Local)
use ;
use LlmAgentBuilder;
use Arc;
async
Supported Models
Google Gemini
| Model | Description |
|---|---|
gemini-3-pro |
Most intelligent model for complex agentic workflows (2M context) |
gemini-3-flash |
Fast and efficient for most tasks (1M context) |
gemini-2.5-pro |
Advanced reasoning and multimodal understanding |
gemini-2.5-flash |
Balanced speed and capability (recommended) |
gemini-2.5-flash-lite |
Ultra-fast for high-volume tasks |
gemini-2.0-flash |
Previous generation (retiring March 2026) |
See Gemini models documentation for the full list.
OpenAI
| Model | Description |
|---|---|
gpt-5.1 |
Latest iteration with improved performance (256K context) |
gpt-5 |
State-of-the-art unified model with adaptive thinking |
gpt-5-mini |
Efficient version for most tasks (128K context) |
gpt-4o |
Multimodal model (deprecated August 2025) |
gpt-4o-mini |
Fast and affordable (deprecated August 2025) |
See OpenAI models documentation for the full list.
Anthropic Claude
| Model | Description |
|---|---|
claude-opus-4.5 |
Most capable model for complex autonomous tasks (200K context) |
claude-sonnet-4.5 |
Balanced intelligence and cost for production |
claude-haiku-4.5 |
Ultra-efficient for high-volume workloads |
claude-opus-4 |
Hybrid model with extended thinking |
claude-sonnet-4 |
Balanced model with extended thinking |
See Anthropic models documentation for the full list.
DeepSeek
| Model | Description |
|---|---|
deepseek-r1-0528 |
Latest reasoning model with enhanced thinking depth (128K context) |
deepseek-r1 |
Advanced reasoning comparable to o1 |
deepseek-v3.1 |
Latest 671B MoE model for general tasks |
deepseek-chat |
671B MoE model, excellent for code (V3) |
deepseek-vl2 |
Vision-language model (32K context) |
Features:
- Thinking Mode - Chain-of-thought reasoning with
<thinking>tags - Context Caching - Automatic KV cache for repeated prefixes (10x cost reduction)
- Tool Calling - Full function calling support
See DeepSeek API documentation for the full list.
Groq
| Model | Description |
|---|---|
llama-4-scout |
Llama 4 Scout (17Bx16E) - Fast via Groq LPU (128K context) |
llama-3.2-90b-text-preview |
Large text model |
llama-3.2-11b-text-preview |
Balanced text model |
llama-3.1-70b-versatile |
Versatile large model |
llama-3.1-8b-instant |
Ultra-fast instruction model |
mixtral-8x7b-32768 |
MoE model with 32K context |
Features:
- Ultra-Fast - LPU-based inference (fastest in the industry)
- Tool Calling - Full function calling support
- Large Context - Up to 128K tokens
See Groq documentation for the full list.
Ollama (Local)
| Model | Description |
|---|---|
llama3.3:70b |
Llama 3.3 70B - Latest for local deployment (128K context) |
llama3.2:3b |
Efficient small model |
llama3.1:8b |
Popular balanced model |
deepseek-r1:14b |
Distilled reasoning model |
deepseek-r1:32b |
Larger distilled reasoning model |
qwen3:14b |
Strong multilingual and coding |
qwen2.5:7b |
Efficient multilingual model (recommended for tool calling) |
mistral:7b |
Fast and capable |
mistral-nemo:12b |
Enhanced Mistral variant (128K context) |
gemma3:9b |
Google's efficient open model |
devstral:24b |
Optimized for coding tasks |
codellama:13b |
Code-focused Llama variant |
Features:
- Local Inference - No API key required
- Privacy - Data stays on your machine
- Tool Calling - Full function calling support (uses non-streaming for reliability)
- MCP Integration - Connect to MCP servers for external tools
See Ollama library for all available models.
Features
- Streaming - Real-time response streaming for all providers
- Tool Calling - Function calling support across all providers
- Async - Full async/await support with backpressure
- Retry - Automatic retry with exponential backoff
- Generation Config - Temperature, top_p, top_k, max_tokens
Environment Variables
# Google Gemini
GOOGLE_API_KEY=your-google-api-key
# OpenAI
OPENAI_API_KEY=your-openai-api-key
# Anthropic
ANTHROPIC_API_KEY=your-anthropic-api-key
# DeepSeek
DEEPSEEK_API_KEY=your-deepseek-api-key
# Groq
GROQ_API_KEY=your-groq-api-key
# Ollama (no key needed, just start the server)
# ollama serve
Feature Flags
Enable specific providers with feature flags:
[]
# All providers (default)
= { = "0.3.1", = ["all-providers"] }
# Individual providers
= { = "0.3.1", = ["gemini"] }
= { = "0.3.1", = ["openai"] }
= { = "0.3.1", = ["anthropic"] }
= { = "0.3.1", = ["deepseek"] }
= { = "0.3.1", = ["groq"] }
= { = "0.3.1", = ["ollama"] }
Related Crates
- adk-rust - Meta-crate with all components
- adk-core - Core
Llmtrait - adk-agent - Agent implementations
License
Apache-2.0
Part of ADK-Rust
This crate is part of the ADK-Rust framework for building AI agents in Rust.