objectiveai 0.1.4

ObjectiveAI SDK, definitions, and utilities
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
//! Reasoning configuration for function executions.

use crate::chat;
use serde::{Deserialize, Serialize};

/// Configuration for generating reasoning summaries during execution.
///
/// When enabled, an LLM summarizes the execution's reasoning process.
#[derive(Debug, Clone, Serialize, Deserialize)]
pub struct Reasoning {
    /// The primary model to use for generating reasoning summaries.
    pub model: chat::completions::request::Model,
    /// Fallback models tried in order if the primary is rate-limited or errors.
    #[serde(skip_serializing_if = "Option::is_none")]
    pub models: Option<Vec<chat::completions::request::Model>>,
}