Skip to main content

vtcode_config/models/model_id/
description.rs

1use super::ModelId;
2
3impl ModelId {
4    /// Get a description of the model's characteristics
5    pub fn description(&self) -> &'static str {
6        if let Some(meta) = self.openrouter_metadata() {
7            return meta.description;
8        }
9        match self {
10            // Gemini models
11            ModelId::Gemini25FlashPreview => {
12                "Latest fast Gemini model with advanced multimodal capabilities"
13            }
14            ModelId::Gemini25Flash => {
15                "Legacy alias for Gemini 2.5 Flash Preview (same capabilities)"
16            }
17            ModelId::Gemini25FlashLite => {
18                "Legacy alias for Gemini 2.5 Flash Preview optimized for efficiency"
19            }
20            ModelId::Gemini25Pro => "Latest most capable Gemini model with reasoning",
21            ModelId::Gemini3ProPreview => {
22                "Preview of next-generation Gemini 3 Pro model with advanced reasoning and capabilities"
23            }
24            // OpenAI models
25            ModelId::GPT5 => "Latest most capable OpenAI model with advanced reasoning",
26            ModelId::GPT5Codex => {
27                "Code-focused GPT-5 variant optimized for tool calling and structured outputs"
28            }
29            ModelId::GPT5Mini => "Latest efficient OpenAI model, great for most tasks",
30            ModelId::GPT5Nano => "Latest most cost-effective OpenAI model",
31            ModelId::CodexMiniLatest => "Latest Codex model optimized for code generation",
32            ModelId::OpenAIGptOss20b => {
33                "OpenAI's open-source 20B parameter GPT-OSS model using harmony tokenization"
34            }
35            ModelId::OpenAIGptOss120b => {
36                "OpenAI's open-source 120B parameter GPT-OSS model using harmony tokenization"
37            }
38            // Anthropic models
39            ModelId::ClaudeOpus45 => {
40                "Latest flagship Anthropic model with exceptional reasoning capabilities"
41            }
42            ModelId::ClaudeOpus41 => {
43                "Latest flagship Anthropic model with exceptional reasoning capabilities"
44            }
45            ModelId::ClaudeSonnet45 => "Latest balanced Anthropic model for general tasks",
46            ModelId::ClaudeHaiku45 => {
47                "Latest efficient Anthropic model optimized for low-latency agent workflows"
48            }
49            ModelId::ClaudeSonnet4 => {
50                "Previous balanced Anthropic model maintained for compatibility"
51            }
52            // DeepSeek models
53            ModelId::DeepSeekChat => {
54                "DeepSeek V3.2 - Fast, efficient chat model for immediate responses"
55            }
56            ModelId::DeepSeekReasoner => {
57                "DeepSeek V3.2 - Thinking mode with integrated tool-use and reasoning capability"
58            }
59            // xAI models
60            ModelId::XaiGrok4 => "Flagship Grok 4 model with long context and tool use",
61            ModelId::XaiGrok4Mini => "Efficient Grok 4 Mini tuned for low latency",
62            ModelId::XaiGrok4Code => "Code-specialized Grok 4 deployment with tool support",
63            ModelId::XaiGrok4CodeLatest => {
64                "Latest Grok 4 code model offering enhanced reasoning traces"
65            }
66            ModelId::XaiGrok4Vision => "Multimodal Grok 4 model with image understanding",
67            // Z.AI models
68            ModelId::ZaiGlm4Plus | ModelId::ZaiGlm4PlusDeepThinking => {
69                "Z.AI flagship model with top-tier capability in reasoning, writing, and tool-use"
70            }
71            ModelId::ZaiGlm47 | ModelId::ZaiGlm47DeepThinking => {
72                "Latest Z.AI GLM flagship with enhanced reasoning, 200k context and coding strengths"
73            }
74            ModelId::ZaiGlm47Flash => {
75                "Z.AI GLM-4.7-Flash 30B-class SOTA lightweight model - Completely free, high-speed, optimized for agentic coding with enhanced reasoning capabilities"
76            }
77            ModelId::ZaiGlm46 | ModelId::ZaiGlm46DeepThinking => {
78                "Previous Z.AI GLM flagship with long-context reasoning and coding strengths"
79            }
80            ModelId::ZaiGlm46V | ModelId::ZaiGlm46VFlash | ModelId::ZaiGlm46VFlashX => {
81                "Vision-capable GLM 4.6 release optimized for multimodal understanding"
82            }
83            ModelId::ZaiGlm45 | ModelId::ZaiGlm45DeepThinking => {
84                "Balanced GLM 4.5 release for general assistant tasks"
85            }
86            ModelId::ZaiGlm45Air => "Efficient GLM 4.5 Air variant tuned for lower latency",
87            ModelId::ZaiGlm45X => "Enhanced GLM 4.5 X variant with improved reasoning",
88            ModelId::ZaiGlm45Airx => "Hybrid GLM 4.5 AirX variant blending efficiency with quality",
89            ModelId::ZaiGlm45Flash => "Low-latency GLM 4.5 Flash optimized for responsiveness",
90            ModelId::ZaiGlm45V => "Vision-capable GLM 4.5 release for multimodal tasks",
91            ModelId::ZaiGlm432b0414128k => {
92                "Legacy GLM 4 32B deployment offering extended 128K context window"
93            }
94            ModelId::OllamaGptOss20b => {
95                "Local GPT-OSS 20B deployment served via Ollama with no external API dependency"
96            }
97            ModelId::OllamaGptOss20bCloud => {
98                "Cloud-hosted GPT-OSS 20B accessed through Ollama Cloud for efficient reasoning tasks"
99            }
100            ModelId::OllamaGptOss120bCloud => {
101                "Cloud-hosted GPT-OSS 120B accessed through Ollama Cloud for larger reasoning tasks"
102            }
103            ModelId::OllamaQwen317b => {
104                "Qwen3 1.7B served locally through Ollama without external API requirements"
105            }
106            ModelId::OllamaDeepseekV32Cloud => {
107                "DeepSeek V3.2 cloud deployment via Ollama with enhanced reasoning and instruction following"
108            }
109            ModelId::OllamaQwen3Next80bCloud => {
110                "Qwen3 Next generation 80B model via Ollama Cloud with improved reasoning and long context"
111            }
112            ModelId::OllamaMistralLarge3675bCloud => {
113                "Mistral Large 3 675B reasoning model via Ollama Cloud for complex problem-solving"
114            }
115            ModelId::OllamaKimiK2ThinkingCloud => {
116                "MoonshotAI Kimi K2 thinking model via Ollama Cloud with explicit reasoning traces"
117            }
118            ModelId::OllamaGlm47Cloud => "Advancing the Coding Capability",
119            ModelId::OllamaMinimaxM21Cloud => {
120                "Exceptional multilingual capabilities to elevate code engineering"
121            }
122            ModelId::OllamaGemini3FlashPreviewCloud => {
123                "Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost."
124            }
125            ModelId::OllamaGemini3ProPreviewLatestCloud => {
126                "Gemini 3 Pro Preview Latest offers advanced reasoning and long context capabilities."
127            }
128            ModelId::OllamaNemotron3Nano30bCloud => {
129                "NVIDIA Nemotron-3-Nano 30B brings efficient excellence to code"
130            }
131            ModelId::OllamaQwen3Coder480bCloud => {
132                "Cloud-hosted Qwen3 Coder 480B model accessed through Ollama Cloud for coding tasks"
133            }
134            ModelId::OllamaGlm46Cloud => {
135                "Cloud-hosted GLM-4.6 model accessed through Ollama Cloud for reasoning and coding"
136            }
137            ModelId::OllamaMinimaxM2Cloud => {
138                "Cloud-hosted MiniMax-M2 model accessed through Ollama Cloud for reasoning tasks"
139            }
140            ModelId::LmStudioMetaLlama38BInstruct => {
141                "Meta Llama 3 8B running through LM Studio's local OpenAI-compatible server"
142            }
143            ModelId::LmStudioMetaLlama318BInstruct => {
144                "Meta Llama 3.1 8B running through LM Studio's local OpenAI-compatible server"
145            }
146            ModelId::LmStudioQwen257BInstruct => {
147                "Qwen2.5 7B hosted in LM Studio for local experimentation and coding tasks"
148            }
149            ModelId::LmStudioGemma22BIt => {
150                "Gemma 2 2B IT deployed via LM Studio for lightweight on-device assistance"
151            }
152            ModelId::LmStudioGemma29BIt => {
153                "Gemma 2 9B IT served locally via LM Studio when you need additional capacity"
154            }
155            ModelId::LmStudioPhi31Mini4kInstruct => {
156                "Phi-3.1 Mini 4K hosted in LM Studio for compact reasoning and experimentation"
157            }
158            ModelId::MinimaxM21 => {
159                "Latest MiniMax-M2.1 model with enhanced code understanding and reasoning"
160            }
161            ModelId::MinimaxM21Lightning => {
162                "Fast version of MiniMax-M2.1 for rapid conversational tasks"
163            }
164            ModelId::MinimaxM2 => {
165                "MiniMax-M2 via Anthropic-compatible API with reasoning and tool use"
166            }
167            ModelId::HuggingFaceDeepseekV32 => {
168                "DeepSeek-V3.2 via Hugging Face router for advanced reasoning"
169            }
170            ModelId::HuggingFaceOpenAIGptOss20b => "OpenAI GPT-OSS 20B via Hugging Face router",
171            ModelId::HuggingFaceOpenAIGptOss120b => "OpenAI GPT-OSS 120B via Hugging Face router",
172            ModelId::HuggingFaceGlm47 => "Z.AI GLM-4.7 via Hugging Face router",
173            ModelId::HuggingFaceGlm47FlashNovita => {
174                "Z.AI GLM-4.7-Flash via Novita inference provider on HuggingFace router. Lightweight model optimized for agentic coding."
175            }
176            ModelId::HuggingFaceKimiK2Thinking => {
177                "MoonshotAI Kimi K2 Thinking via Hugging Face router"
178            }
179            ModelId::HuggingFaceMinimaxM21Novita => {
180                "MiniMax-M2.1 model via Novita inference provider on HuggingFace router. Enhanced reasoning capabilities."
181            }
182            ModelId::HuggingFaceDeepseekV32Novita => {
183                "DeepSeek-V3.2 via Novita inference provider on HuggingFace router."
184            }
185            ModelId::HuggingFaceXiaomiMimoV2FlashNovita => {
186                "Xiaomi MiMo-V2-Flash via Novita on HuggingFace router."
187            }
188            _ => unreachable!(),
189        }
190    }
191}