use std::io;
use std::process::Stdio;
use std::str;
use crate::cli_common::{CliRunnerBase, MAX_OUTPUT_BYTES};
use crate::types::{
ChatRequest, ChatResponse, ChatStream, LlmCapabilities, LlmProvider, RunnerError, StreamChunk,
TokenUsage,
};
use async_trait::async_trait;
use serde::Deserialize;
use tokio::io::{AsyncBufReadExt, BufReader};
use tokio::process::Command;
use tokio_stream::wrappers::LinesStream;
use tokio_stream::StreamExt;
use tracing::instrument;
use crate::config::RunnerConfig;
use crate::process::{read_stderr_capped, run_cli_command};
use crate::prompt::prepare_user_prompt;
use crate::sandbox::{apply_sandbox, build_policy};
use crate::stream::{GuardedStream, MAX_STREAMING_STDERR_BYTES};
#[derive(Debug, Deserialize)]
struct GeminiResponse {
response: Option<String>,
session_id: Option<String>,
#[serde(default)]
stats: Option<GeminiStats>,
}
#[derive(Debug, Deserialize)]
#[allow(clippy::struct_field_names)]
struct GeminiStats {
#[serde(default)]
total_tokens: Option<u32>,
#[serde(default)]
input_tokens: Option<u32>,
#[serde(default)]
output_tokens: Option<u32>,
}
const DEFAULT_MODEL: &str = "gemini-2.5-flash";
const FALLBACK_MODELS: &[&str] = &["gemini-2.5-flash", "gemini-2.5-pro", "gemini-2.0-flash"];
pub struct GeminiCliRunner {
base: CliRunnerBase,
}
impl GeminiCliRunner {
#[must_use]
pub fn new(config: RunnerConfig) -> Self {
Self {
base: CliRunnerBase::new(config, DEFAULT_MODEL, FALLBACK_MODELS),
}
}
pub async fn set_session(&self, key: &str, session_id: &str) {
self.base.set_session(key, session_id).await;
}
fn build_command(&self, prompt: &str, output_format: &str) -> Command {
let mut cmd = Command::new(&self.base.config.binary_path);
cmd.args(["-p", prompt, "-o", output_format]);
cmd.arg("-y");
let model = self
.base
.config
.model
.as_deref()
.unwrap_or_else(|| self.base.default_model());
cmd.args(["-m", model]);
for arg in &self.base.config.extra_args {
cmd.arg(arg);
}
if let Ok(policy) = build_policy(
self.base.config.working_directory.as_deref(),
&self.base.config.allowed_env_keys,
) {
apply_sandbox(&mut cmd, &policy);
}
cmd
}
fn parse_jsonl_response(raw: &[u8]) -> Result<(ChatResponse, Option<String>), RunnerError> {
let text = str::from_utf8(raw).map_err(|e| {
RunnerError::internal(format!("Gemini CLI output is not valid UTF-8: {e}"))
})?;
if let Ok(parsed) = serde_json::from_str::<GeminiResponse>(text) {
let content = parsed.response.unwrap_or_default();
let usage = parsed.stats.map(|s| {
let input = s.input_tokens.unwrap_or(0);
let output = s.output_tokens.unwrap_or(0);
let total = s.total_tokens.unwrap_or(input + output);
TokenUsage {
prompt_tokens: input,
completion_tokens: output,
total_tokens: total,
}
});
return Ok((
ChatResponse {
content,
model: "gemini".to_owned(),
usage,
finish_reason: Some("stop".to_owned()),
warnings: None,
tool_calls: None,
},
parsed.session_id,
));
}
let mut content_parts: Vec<String> = Vec::new();
let mut session_id: Option<String> = None;
let mut usage: Option<TokenUsage> = None;
for line in text.lines() {
let trimmed = line.trim();
if trimmed.is_empty() {
continue;
}
let value: serde_json::Value = match serde_json::from_str(trimmed) {
Ok(v) => v,
Err(_) => continue,
};
let line_type = value.get("type").and_then(|v| v.as_str()).unwrap_or("");
match line_type {
"init" => {
if let Some(sid) = value.get("session_id").and_then(|v| v.as_str()) {
session_id = Some(sid.to_owned());
}
}
"message" => {
let role = value.get("role").and_then(|v| v.as_str()).unwrap_or("");
if role == "assistant" {
if let Some(c) = value.get("content").and_then(|v| v.as_str()) {
content_parts.push(c.to_owned());
}
}
}
"result" => {
if let Some(stats) = value.get("stats") {
let input = stats
.get("input_tokens")
.and_then(serde_json::Value::as_u64)
.unwrap_or(0);
let output = stats
.get("output_tokens")
.and_then(serde_json::Value::as_u64)
.unwrap_or(0);
let total = stats
.get("total_tokens")
.and_then(serde_json::Value::as_u64)
.unwrap_or(input + output);
#[allow(clippy::cast_possible_truncation)]
{
usage = Some(TokenUsage {
prompt_tokens: input as u32,
completion_tokens: output as u32,
total_tokens: total as u32,
});
}
}
}
_ => {}
}
}
let content = content_parts.join("");
Ok((
ChatResponse {
content,
model: "gemini".to_owned(),
usage,
finish_reason: Some("stop".to_owned()),
warnings: None,
tool_calls: None,
},
session_id,
))
}
}
#[async_trait]
impl LlmProvider for GeminiCliRunner {
crate::delegate_provider_base!(
"gemini",
"Gemini CLI",
LlmCapabilities::STREAMING | LlmCapabilities::TEMPERATURE | LlmCapabilities::MAX_TOKENS
);
#[instrument(skip_all, fields(runner = "gemini"))]
async fn complete(&self, request: &ChatRequest) -> Result<ChatResponse, RunnerError> {
let prepared = prepare_user_prompt(&request.messages)?;
let prompt = &prepared.prompt;
let mut cmd = self.build_command(prompt, "json");
if let Some(model) = &request.model {
if let Some(sid) = self.base.get_session(model).await {
cmd.args(["--resume", &sid]);
}
}
let output = run_cli_command(&mut cmd, self.base.config.timeout, MAX_OUTPUT_BYTES).await?;
self.base.check_exit_code(&output, "gemini")?;
let (response, session_id) = Self::parse_jsonl_response(&output.stdout)?;
if let Some(sid) = session_id {
if let Some(model) = &request.model {
self.base.set_session(model, &sid).await;
}
}
Ok(response)
}
#[instrument(skip_all, fields(runner = "gemini"))]
async fn complete_stream(&self, request: &ChatRequest) -> Result<ChatStream, RunnerError> {
let prepared = prepare_user_prompt(&request.messages)?;
let prompt = &prepared.prompt;
let mut cmd = self.build_command(prompt, "stream-json");
if let Some(model) = &request.model {
if let Some(sid) = self.base.get_session(model).await {
cmd.args(["--resume", &sid]);
}
}
cmd.stdout(Stdio::piped());
cmd.stderr(Stdio::piped());
let mut child = cmd.spawn().map_err(|e| {
RunnerError::internal(format!("Failed to spawn gemini for streaming: {e}"))
})?;
let stdout = child.stdout.take().ok_or_else(|| {
RunnerError::internal("Failed to capture gemini stdout for streaming")
})?;
let stderr_task = tokio::spawn(read_stderr_capped(
child.stderr.take(),
MAX_STREAMING_STDERR_BYTES,
));
let reader = BufReader::new(stdout);
let lines = LinesStream::new(reader.lines());
let stream = lines.map(move |line_result: Result<String, io::Error>| {
let line = line_result
.map_err(|e| RunnerError::internal(format!("Error reading gemini stream: {e}")))?;
if line.trim().is_empty() {
return Ok(StreamChunk {
delta: String::new(),
is_final: false,
finish_reason: None,
});
}
let value: serde_json::Value = serde_json::from_str(&line).map_err(|e| {
RunnerError::internal(format!("Invalid JSON in gemini stream: {e}"))
})?;
let chunk_type = value.get("type").and_then(|v| v.as_str()).unwrap_or("");
match chunk_type {
"message" => {
let role = value.get("role").and_then(|v| v.as_str()).unwrap_or("");
if role == "assistant" {
Ok(StreamChunk {
delta: value
.get("content")
.and_then(|v| v.as_str())
.unwrap_or("")
.to_owned(),
is_final: false,
finish_reason: None,
})
} else {
Ok(StreamChunk {
delta: String::new(),
is_final: false,
finish_reason: None,
})
}
}
"result" => Ok(StreamChunk {
delta: String::new(),
is_final: true,
finish_reason: Some("stop".to_owned()),
}),
_ => Ok(StreamChunk {
delta: String::new(),
is_final: false,
finish_reason: None,
}),
}
});
Ok(Box::pin(GuardedStream::new(stream, child, stderr_task)))
}
}
#[cfg(test)]
mod tests {
use super::*;
use std::path::PathBuf;
#[test]
fn test_parse_single_json_response() {
let json = br#"{"session_id":"abc123","response":"hello from gemini","stats":{"input_tokens":10,"output_tokens":5,"total_tokens":15}}"#;
let (resp, sid) = GeminiCliRunner::parse_jsonl_response(json).unwrap();
assert_eq!(resp.content, "hello from gemini");
assert_eq!(sid, Some("abc123".to_owned()));
let usage = resp.usage.unwrap();
assert_eq!(usage.prompt_tokens, 10);
assert_eq!(usage.completion_tokens, 5);
assert_eq!(usage.total_tokens, 15);
}
#[test]
fn test_parse_single_json_response_no_stats() {
let json = br#"{"response":"hello"}"#;
let (resp, sid) = GeminiCliRunner::parse_jsonl_response(json).unwrap();
assert_eq!(resp.content, "hello");
assert!(sid.is_none());
assert!(resp.usage.is_none());
}
#[test]
fn test_parse_jsonl_response() {
let jsonl = b"
{\"type\":\"init\",\"session_id\":\"sess-42\",\"model\":\"auto-gemini-3\"}
{\"type\":\"message\",\"role\":\"user\",\"content\":\"hi\"}
{\"type\":\"message\",\"role\":\"assistant\",\"content\":\"hello from gemini\",\"delta\":true}
{\"type\":\"result\",\"status\":\"success\",\"stats\":{\"total_tokens\":8628,\"input_tokens\":100,\"output_tokens\":50}}
";
let (resp, sid) = GeminiCliRunner::parse_jsonl_response(jsonl).unwrap();
assert_eq!(resp.content, "hello from gemini");
assert_eq!(sid, Some("sess-42".to_owned()));
let usage = resp.usage.unwrap();
assert_eq!(usage.prompt_tokens, 100);
assert_eq!(usage.completion_tokens, 50);
assert_eq!(usage.total_tokens, 8628);
}
#[test]
fn test_default_model() {
let config = RunnerConfig::new(PathBuf::from("gemini"));
let runner = GeminiCliRunner::new(config);
assert_eq!(runner.default_model(), "gemini-2.5-flash");
}
#[test]
fn test_custom_model() {
let config = RunnerConfig::new(PathBuf::from("gemini")).with_model("gemini-2.5-pro");
let runner = GeminiCliRunner::new(config);
assert_eq!(runner.default_model(), "gemini-2.5-pro");
}
#[test]
fn test_available_models() {
let config = RunnerConfig::new(PathBuf::from("gemini"));
let runner = GeminiCliRunner::new(config);
let models = runner.available_models();
assert_eq!(models.len(), 3);
assert!(models.contains(&"gemini-2.5-flash".to_owned()));
assert!(models.contains(&"gemini-2.5-pro".to_owned()));
assert!(models.contains(&"gemini-2.0-flash".to_owned()));
}
#[test]
fn test_capabilities() {
let config = RunnerConfig::new(PathBuf::from("gemini"));
let runner = GeminiCliRunner::new(config);
assert!(runner.capabilities().supports_streaming());
}
#[test]
fn test_name_and_display() {
let config = RunnerConfig::new(PathBuf::from("gemini"));
let runner = GeminiCliRunner::new(config);
assert_eq!(runner.name(), "gemini");
assert_eq!(runner.display_name(), "Gemini CLI");
}
}