Skip to main content

vtcode_config/models/model_id/
description.rs

1use super::ModelId;
2
3impl ModelId {
4    /// Get a description of the model's characteristics
5    pub fn description(&self) -> &'static str {
6        if let Some(meta) = self.openrouter_metadata() {
7            return meta.description;
8        }
9        match self {
10            // Gemini models
11            ModelId::Gemini3ProPreview => {
12                "Preview of next-generation Gemini 3 Pro model with advanced reasoning and capabilities"
13            }
14            ModelId::Gemini3FlashPreview => {
15                "Our most intelligent model built for speed, combining frontier intelligence with superior search and grounding"
16            }
17            // OpenAI models
18            ModelId::GPT5 => "Latest most capable OpenAI model with advanced reasoning",
19            ModelId::GPT52 => {
20                "Latest flagship OpenAI model with improved reasoning, xhigh effort, and built-in compaction support"
21            }
22            ModelId::GPT52Codex => {
23                "GPT-5.2 variant optimized for agentic coding tasks with reasoning effort support"
24            }
25            ModelId::GPT5Codex => {
26                "Code-focused GPT-5 variant optimized for tool calling and structured outputs"
27            }
28            ModelId::GPT5Mini => "Latest efficient OpenAI model, great for most tasks",
29            ModelId::GPT5Nano => "Latest most cost-effective OpenAI model",
30            ModelId::GPT51 => {
31                "Enhanced most capable OpenAI model with improved reasoning and capabilities"
32            }
33            ModelId::GPT51Codex => {
34                "Code-focused GPT-5.1 variant optimized for tool calling and structured outputs"
35            }
36            ModelId::GPT51CodexMax => {
37                "Maximum context code-focused GPT-5.1 variant optimized for large codebases"
38            }
39            ModelId::GPT51Mini => "Enhanced efficient OpenAI model with improved capabilities",
40            ModelId::CodexMiniLatest => "Latest Codex model optimized for code generation",
41            ModelId::OpenAIGptOss20b => {
42                "OpenAI's open-source 20B parameter GPT-OSS model using harmony tokenization"
43            }
44            ModelId::OpenAIGptOss120b => {
45                "OpenAI's open-source 120B parameter GPT-OSS model using harmony tokenization"
46            }
47            // Anthropic models
48            ModelId::ClaudeOpus46 => {
49                "Next-gen Anthropic flagship with extended and adaptive thinking support"
50            }
51            ModelId::ClaudeOpus45 => {
52                "Latest flagship Anthropic model with exceptional reasoning capabilities"
53            }
54            ModelId::ClaudeOpus41 => {
55                "Latest flagship Anthropic model with exceptional reasoning capabilities"
56            }
57            ModelId::ClaudeSonnet45 => "Latest balanced Anthropic model for general tasks",
58            ModelId::ClaudeHaiku45 => {
59                "Latest efficient Anthropic model optimized for low-latency agent workflows"
60            }
61            ModelId::ClaudeSonnet4 => {
62                "Previous balanced Anthropic model maintained for compatibility"
63            }
64            ModelId::ClaudeOpus4 => "Previous generation premium flagship model",
65            ModelId::ClaudeSonnet37 => "Latest model in the Claude 3 family with extended thinking",
66            ModelId::ClaudeHaiku35 => "Highly efficient model for high-volume tasks",
67            // DeepSeek models
68            ModelId::DeepSeekChat => {
69                "DeepSeek V3.2 - Fast, efficient chat model for immediate responses"
70            }
71            ModelId::DeepSeekReasoner => {
72                "DeepSeek V3.2 - Thinking mode with integrated tool-use and reasoning capability"
73            }
74            // xAI models
75            ModelId::XaiGrok4 => "Flagship Grok 4 model with long context and tool use",
76            ModelId::XaiGrok4Mini => "Efficient Grok 4 Mini tuned for low latency",
77            ModelId::XaiGrok4Code => "Code-specialized Grok 4 deployment with tool support",
78            ModelId::XaiGrok4CodeLatest => {
79                "Latest Grok 4 code model offering enhanced reasoning traces"
80            }
81            ModelId::XaiGrok4Vision => "Multimodal Grok 4 model with image understanding",
82            // Z.AI models
83            ModelId::ZaiGlm5 => {
84                "Z.ai flagship GLM-5 foundation model engineered for complex systems design and long-horizon agent workflows"
85            }
86            // Ollama models
87            ModelId::OllamaGptOss20b => {
88                "Local GPT-OSS 20B deployment served via Ollama with no external API dependency"
89            }
90            ModelId::OllamaGptOss20bCloud => {
91                "Cloud-hosted GPT-OSS 20B accessed through Ollama Cloud for efficient reasoning tasks"
92            }
93            ModelId::OllamaGptOss120bCloud => {
94                "Cloud-hosted GPT-OSS 120B accessed through Ollama Cloud for larger reasoning tasks"
95            }
96            ModelId::OllamaQwen317b => {
97                "Qwen3 1.7B served locally through Ollama without external API requirements"
98            }
99            ModelId::OllamaDeepseekV32Cloud => {
100                "DeepSeek V3.2 cloud deployment via Ollama with enhanced reasoning and instruction following"
101            }
102            ModelId::OllamaQwen3Next80bCloud => {
103                "Qwen3 Next generation 80B model via Ollama Cloud with improved reasoning and long context"
104            }
105            ModelId::OllamaMistralLarge3675bCloud => {
106                "Mistral Large 3 675B reasoning model via Ollama Cloud for complex problem-solving"
107            }
108            ModelId::OllamaGlm5Cloud => "Cloud-hosted GLM-5 model served via Ollama Cloud",
109            ModelId::OllamaMinimaxM25Cloud => {
110                "Exceptional multilingual capabilities to elevate code engineering"
111            }
112            ModelId::OllamaGemini3FlashPreviewCloud => {
113                "Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost."
114            }
115            ModelId::OllamaGemini3ProPreviewLatestCloud => {
116                "Gemini 3 Pro Preview Latest offers advanced reasoning and long context capabilities."
117            }
118            ModelId::OllamaNemotron3Nano30bCloud => {
119                "NVIDIA Nemotron-3-Nano 30B brings efficient excellence to code"
120            }
121            ModelId::OllamaQwen3Coder480bCloud => {
122                "Cloud-hosted Qwen3 Coder 480B model accessed through Ollama Cloud for coding tasks"
123            }
124            ModelId::OllamaMinimaxM2Cloud => {
125                "Cloud-hosted MiniMax-M2 model accessed through Ollama Cloud for reasoning tasks"
126            }
127            ModelId::LmStudioMetaLlama38BInstruct => {
128                "Meta Llama 3 8B running through LM Studio's local OpenAI-compatible server"
129            }
130            ModelId::LmStudioMetaLlama318BInstruct => {
131                "Meta Llama 3.1 8B running through LM Studio's local OpenAI-compatible server"
132            }
133            ModelId::LmStudioQwen257BInstruct => {
134                "Qwen2.5 7B hosted in LM Studio for local experimentation and coding tasks"
135            }
136            ModelId::LmStudioGemma22BIt => {
137                "Gemma 2 2B IT deployed via LM Studio for lightweight on-device assistance"
138            }
139            ModelId::LmStudioGemma29BIt => {
140                "Gemma 2 9B IT served locally via LM Studio when you need additional capacity"
141            }
142            ModelId::LmStudioPhi31Mini4kInstruct => {
143                "Phi-3.1 Mini 4K hosted in LM Studio for compact reasoning and experimentation"
144            }
145            ModelId::MinimaxM25 => {
146                "Latest MiniMax-M2.5 model with further improvements in reasoning and coding"
147            }
148            ModelId::MinimaxM2 => {
149                "MiniMax-M2 via Anthropic-compatible API with reasoning and tool use"
150            }
151            ModelId::MoonshotMinimaxM25 => "MiniMax-M2.5 served via Moonshot API",
152            ModelId::MoonshotQwen3CoderNext => "Qwen3 Coder Next model served via Moonshot API",
153            ModelId::HuggingFaceDeepseekV32 => {
154                "DeepSeek-V3.2 via Hugging Face router for advanced reasoning"
155            }
156            ModelId::HuggingFaceOpenAIGptOss20b => "OpenAI GPT-OSS 20B via Hugging Face router",
157            ModelId::HuggingFaceOpenAIGptOss120b => "OpenAI GPT-OSS 120B via Hugging Face router",
158            ModelId::HuggingFaceMinimaxM25Novita => {
159                "MiniMax-M2.5 model via Novita inference provider on HuggingFace router. Enhanced reasoning capabilities."
160            }
161            ModelId::HuggingFaceDeepseekV32Novita => {
162                "DeepSeek-V3.2 via Novita inference provider on HuggingFace router."
163            }
164            ModelId::HuggingFaceXiaomiMimoV2FlashNovita => {
165                "Xiaomi MiMo-V2-Flash via Novita on HuggingFace router."
166            }
167            ModelId::HuggingFaceGlm5Novita => {
168                "Z.ai GLM-5 flagship model via Novita inference provider on HuggingFace router."
169            }
170            ModelId::HuggingFaceQwen3CoderNextNovita => {
171                "Qwen3-Coder-Next via Novita inference provider on HuggingFace router. Coding-optimized model with reasoning capabilities."
172            }
173            ModelId::OpenRouterMinimaxM25 => "MiniMax-M2.5 flagship model via OpenRouter",
174            ModelId::OpenRouterQwen3CoderNext => {
175                "Next-generation Qwen3 coding model optimized for agentic workflows via OpenRouter"
176            }
177            _ => unreachable!(),
178        }
179    }
180}