gemini-rust
A comprehensive Rust client library for Google's Gemini 2.5 API.
✨ Features
- 🚀 Complete Gemini 2.5 API Implementation - Full support for all Gemini API endpoints
- 🛠️ Function Calling & Tools - Custom functions and Google Search integration with OpenAPI schema support
- 📦 Batch Processing - Efficient batch content generation and embedding
- 💾 Content Caching - Cache system instructions and conversation history for cost optimization
- 🔄 Streaming Responses - Real-time streaming of generated content
- 🧠 Thinking Mode - Support for Gemini 2.5 thinking capabilities
- 🎨 Image Generation - Text-to-image generation and image editing capabilities
- 🎤 Speech Generation - Text-to-speech with single and multi-speaker support
- 🖼️ Multimodal Support - Images and binary data processing
- 📊 Text Embeddings - Advanced embedding generation with multiple task types
- ⚙️ Highly Configurable - Custom models, endpoints, and generation parameters with HTTP client builder
- 🔒 Type Safe - Comprehensive type definitions with full
serde
support - ⚡ Async/Await - Built on
tokio
for high-performance async operations - 🔍 Comprehensive Tracing - Built-in structured logging and telemetry with
tracing
for observability
📦 Installation
Add this to your Cargo.toml
:
[]
= "1.5.1"
🚀 Quick Start
Basic Content Generation
Get started with simple text generation, system prompts, and conversations. See basic_generation.rs
for complete examples including simple messages, system prompts, and multi-turn conversations.
Streaming Responses
Enable real-time content streaming for interactive applications. See basic_streaming.rs
for examples of processing content as it's generated with immediate display.
🛠️ Key Features
The library provides comprehensive access to all Gemini 2.5 capabilities through an intuitive Rust API:
🧠 Thinking Mode (Gemini 2.5)
Advanced reasoning capabilities with thought process visibility and custom thinking budgets. See thinking_basic.rs
and thinking_advanced.rs
.
🛠️ Function Calling & Tools
- Custom function declarations with OpenAPI schema support (using
schemars
) - Google Search integration for real-time information
- Type-safe function definitions with automatic schema generation
- See
tools.rs
andcomplex_function.rs
🎨 Multimodal Generation
- Image Generation: Text-to-image with detailed prompts and editing capabilities
- Speech Generation: Text-to-speech with single and multi-speaker support
- Image Processing: Analyze images, videos, and binary data
- See
image_generation.rs
andmulti_speaker_tts.rs
📦 Batch Processing
Efficient processing of multiple requests with automatic file handling for large jobs. See batch_generate.rs
.
💾 Content Caching
Cache system instructions and conversation history to reduce costs and improve performance. See cache_basic.rs
.
📊 Text Embeddings
Advanced embedding generation with multiple task types for document retrieval and semantic search. See embedding.rs
.
🔄 Streaming Responses
Real-time streaming of generated content for interactive applications. See streaming.rs
.
⚙️ Highly Configurable
- Custom models and endpoints
- Detailed generation parameters (temperature, tokens, etc.)
- HTTP client customization with timeouts and proxies
- See
generation_config.rs
andcustom_base_url.rs
🔍 Observability
Built-in structured logging and telemetry with tracing
for comprehensive monitoring and debugging.
🔧 Configuration
Custom Models
Configure different Gemini models including Flash, Pro, Lite, and custom models. See custom_models.rs
for examples of all model configuration options including convenience methods, enum variants, and custom model strings.
Custom Base URL
Use custom API endpoints and configurations. See custom_base_url.rs
for examples of configuring custom endpoints with different models.
Configurable HTTP Client Builder
For advanced HTTP configuration (timeouts, proxies, custom headers), use the builder pattern. See http_client_builder.rs
for a complete example with custom timeouts, user agents, connection pooling, and proxy configuration.
🔍 Tracing and Telemetry
The library is instrumented with the tracing
crate to provide detailed telemetry data for monitoring and debugging. This allows you to gain deep insights into the library's performance and behavior.
Key tracing features include:
- HTTP Request Tracing: Captures detailed information about every API call, including HTTP method, URL, and response status, to help diagnose network-related issues
- Token Usage Monitoring: Records the number of prompt, candidate, and total tokens for each generation request, enabling cost analysis and optimization
- Structured Logging: Emits traces as structured events, compatible with modern log aggregation platforms like Elasticsearch, Datadog, and Honeycomb, allowing for powerful querying and visualization
- Performance Metrics: Provides timing information for each API request, allowing you to identify and address performance bottlenecks
To use these features, you will need to integrate a tracing
subscriber into your application. See tracing_telemetry.rs
for comprehensive examples including basic console logging, structured logging for production, and environment-based log level filtering.
📚 Examples
The repository includes 30+ comprehensive examples demonstrating all features. See examples/README.md
for detailed information.
Quick Start Examples
basic_generation.rs
- Simple content generation for beginnersbasic_streaming.rs
- Real-time streaming responsessimple.rs
- Comprehensive example with function callingthinking_basic.rs
- Gemini 2.5 thinking modebatch_generate.rs
- Batch content generationimage_generation.rs
- Text-to-image generationgoogle_search.rs
- Google Search integrationurl_context.rs
- URL Context tool for web content analysis
Run any example:
GEMINI_API_KEY="your-api-key"
🔑 API Key Setup
Get your API key from Google AI Studio and set it as an environment variable:
🚦 Supported Models
- Gemini 2.5 Flash - Fast, efficient model (default) -
Model::Gemini25Flash
- Gemini 2.5 Flash Lite - Lightweight model -
Model::Gemini25FlashLite
- Gemini 2.5 Pro - Advanced model with thinking capabilities -
Model::Gemini25Pro
- Text Embedding 004 - Latest embedding model -
Model::TextEmbedding004
- Custom models - Use
Model::Custom(String)
or string literals for other models
🤝 Contributing
Contributions are welcome! Please feel free to submit a Pull Request. For major changes, please open an issue first to discuss what you would like to change.
For guidelines on developing agents and applications, see the Agent Development Guide.
📄 License
This project is licensed under the MIT License - see the LICENSE file for details.
🙏 Acknowledgments
- Google for providing the Gemini API
- The Rust community for excellent async and HTTP libraries
- Special thanks to @npatsakula for major contributions that made this project more complete
- All contributors who have helped improve this library