gemini-rust
A comprehensive Rust client library for Google's Gemini API.
β¨ Features
- π Complete Gemini API Implementation - Full support for all Gemini API endpoints
- π οΈ Function Calling & Tools - Custom functions, Google Search, and Google Maps integration with OpenAPI schema support
- πΊοΈ Google Maps Grounding - Location-aware responses with Google Maps data and widget support
- π¦ Batch Processing - Efficient batch content generation and embedding
- πΎ Content Caching - Cache system instructions and conversation history for cost optimization
- π Streaming Responses - Real-time streaming of generated content
- π§ Thinking Mode - Support for Gemini 2.5+ thinking capabilities
- π Gemini 3 Pro - Code execution, advanced thinking levels, and media resolution control
- π’ Token Count API - Pre-calculate token usage for cost optimization
- π― Safety Settings - Customize content moderation and safety filters
- π File Handles - Efficient file reference without re-encoding large files
- π Image Generation - Text-to-image generation and image editing capabilities
- π€ Speech Generation - Text-to-speech with single and multi-speaker support
- πΌοΈ Multimodal Support - Images and binary data processing
- π Text Embeddings - Advanced embedding generation with multiple task types
- βοΈ Highly Configurable - Custom models, endpoints, and generation parameters with HTTP client builder
- π Type Safe - Comprehensive type definitions with full
serdesupport - β‘ Async/Await - Built on
tokiofor high-performance async operations - π Comprehensive Tracing - Built-in structured logging and telemetry with
tracingfor observability
π¦ Installation
Add this to your Cargo.toml:
[]
= "1.6.1"
π Quick Start
Basic Content Generation
Get started with simple text generation, system prompts, and conversations. See basic_generation.rs for complete examples including simple messages, system prompts, and multi-turn conversations.
Streaming Responses
Enable real-time content streaming for interactive applications. See basic_streaming.rs for examples of processing content as it's generated with immediate display.
Google Maps Grounding
Add location-aware capabilities to your applications with Google Maps integration. See simple_maps_example.rs for basic usage and google_maps_grounding.rs for comprehensive examples.
Token Count API
Calculate token usage before making generation requests for cost estimation and optimization. See count_tokens.rs.
Safety Settings
Customize content moderation with granular control over different harm categories and block thresholds. See safety_settings.rs.
Gemini 3 Pro
Access the latest model features including code execution and advanced thinking levels. See gemini_3_code_execution.rs for code execution and gemini_3_thinking_and_media.rs for thinking levels.
π οΈ Key Features
The library provides comprehensive access to all Gemini 2.5 capabilities through an intuitive Rust API:
π§ Thinking Mode (Gemini 2.5)
Advanced reasoning capabilities with thought process visibility and custom thinking budgets. See thinking_basic.rs and thinking_advanced.rs.
π οΈ Function Calling & Tools
- Custom function declarations with OpenAPI schema support (using
schemars) - Google Search integration for real-time information
- Google Maps grounding for location-aware responses
- Type-safe function definitions with automatic schema generation
- See
tools.rs,complex_function.rs, andgoogle_maps_grounding.rs
πΊοΈ Google Maps Grounding
- Location-Aware Responses: Access Google Maps data for geographically specific queries
- Widget Support: Generate context tokens for interactive Google Maps widgets
- Grounding Sources: Access citation information for all Maps data used in responses
- Easy Integration: Simple API with location context configuration
- See
simple_maps_example.rsandadvanced_maps_configuration.rs
π¨ Multimodal Generation
- Image Generation: Nano Banana (Flash) and Pro (Gemini 3) Text-to-image with detailed thinking and follow up editing capabilities
- Speech Generation: Text-to-speech with single and multi-speaker support
- Image Processing: Analyze images, videos, and binary data
- See
image_generation.rsandmulti_speaker_tts.rs
π¦ Batch Processing
Efficient processing of multiple requests with automatic file handling for large jobs. See batch_generate.rs.
πΎ Content Caching
Cache system instructions and conversation history to reduce costs and improve performance. See cache_basic.rs.
π Text Embeddings
Advanced embedding generation with multiple task types for document retrieval and semantic search. See embedding.rs.
π Streaming Responses
Real-time streaming of generated content for interactive applications. See streaming.rs.
βοΈ Highly Configurable
- Custom models and endpoints
- Detailed generation parameters (temperature, tokens, etc.)
- HTTP client customization with timeouts and proxies
- See
generation_config.rsandcustom_base_url.rs
π Observability
Built-in structured logging and telemetry with tracing for comprehensive monitoring and debugging.
π’ Token Count API
Pre-calculate token usage for cost estimation and optimization. Calculate tokens for your requests before executing them. See count_tokens.rs.
π― Safety Settings
Customize content moderation with granular control over different harm categories (Hate Speech, Dangerous Content, etc.) and block thresholds (Block None, Low, Medium, High). See safety_settings.rs.
π Gemini 3 Pro
- Code Execution: Generate and execute Python code for mathematical calculations, data analysis, and computational tasks
- Thinking Levels: Choose Low for faster responses or High for deeper analysis
- Media Resolution: Fine-grained control over image and PDF processing quality
- See
gemini_3_code_execution.rsandgemini_3_thinking_and_media.rs
π File Handles
Efficiently reference previously uploaded files without re-encoding. Upload files once and reference them multiple times, reducing data transfer. Supports PDFs, images, and other binary formats. See file_input.rs and files_usage.rs.
π§ Configuration
Custom Models
Configure different Gemini models including Flash, Pro, Lite, and custom models. See custom_models.rs for examples of all model configuration options including convenience methods, enum variants, and custom model strings.
Custom Base URL
Use custom API endpoints and configurations. See custom_base_url.rs for examples of configuring custom endpoints with different models.
Configurable HTTP Client Builder
For advanced HTTP configuration (timeouts, proxies, custom headers), use the builder pattern. See http_client_builder.rs for a complete example with custom timeouts, user agents, connection pooling, and proxy configuration.
π Tracing and Telemetry
The library is instrumented with the tracing crate to provide detailed telemetry data for monitoring and debugging. This allows you to gain deep insights into the library's performance and behavior.
Key tracing features include:
- HTTP Request Tracing: Captures detailed information about every API call, including HTTP method, URL, and response status, to help diagnose network-related issues
- Token Usage Monitoring: Records the number of prompt, candidate, and total tokens for each generation request, enabling cost analysis and optimization
- Structured Logging: Emits traces as structured events, compatible with modern log aggregation platforms like Elasticsearch, Datadog, and Honeycomb, allowing for powerful querying and visualization
- Performance Metrics: Provides timing information for each API request, allowing you to identify and address performance bottlenecks
To use these features, you will need to integrate a tracing subscriber into your application. See tracing_telemetry.rs for comprehensive examples including basic console logging, structured logging for production, and environment-based log level filtering.
π Examples
The repository includes 30+ comprehensive examples demonstrating all features. See examples/README.md for detailed information.
Quick Start Examples
basic_generation.rs- Simple content generation for beginnersbasic_streaming.rs- Real-time streaming responsessimple.rs- Comprehensive example with function callingthinking_basic.rs- Gemini 2.5 thinking modecount_tokens.rs- Pre-calculate token usagesafety_settings.rs- Configure safety filtersgemini_3_code_execution.rs- Code execution with Pythonfile_input.rs- Upload and reference filesbatch_generate.rs- Batch content generationimage_generation.rs- Text-to-image generationgoogle_search.rs- Google Search integrationurl_context.rs- URL Context tool for web content analysis
Run any example:
GEMINI_API_KEY="your-api-key"
π API Key Setup
Get your API key from Google AI Studio and set it as an environment variable:
π¦ Supported Models
- Gemini 2.5 Flash - Fast, efficient model (default) -
Model::Gemini25Flash - Gemini 2.5 Flash Lite - Lightweight model -
Model::Gemini25FlashLite - Gemini 2.5 Pro - Advanced model with thinking capabilities -
Model::Gemini25Pro - Gemini 3 Pro - Latest model with code execution and advanced thinking -
Model::Gemini3Pro(Preview) - Text Embedding 004 - Latest embedding model -
Model::TextEmbedding004 - Custom models - Use
Model::Custom(String)or string literals for other models
π€ Contributing
Contributions are welcome! Please feel free to submit a Pull Request. For major changes, please open an issue first to discuss what you would like to change.
For guidelines on developing agents and applications, see the Agent Development Guide.
π License
This project is licensed under the MIT License - see the LICENSE file for details.
π Acknowledgments
- Google for providing the Gemini API
- The Rust community for excellent async and HTTP libraries
- Special thanks to @npatsakula for major contributions that made this project more complete
- All contributors who have helped improve this library