vtcode_config/models/model_id/
description.rs1use super::ModelId;
2
3impl ModelId {
4 pub fn description(&self) -> &'static str {
6 if let Some(meta) = self.openrouter_metadata() {
7 return meta.description;
8 }
9 match self {
10 ModelId::Gemini31ProPreview => {
12 "Latest Gemini 3.1 Pro flagship model with improved thinking, efficiency, and factual consistency"
13 }
14 ModelId::Gemini31ProPreviewCustomTools => {
15 "Gemini 3.1 Pro variant optimized for agentic workflows using custom tools and bash"
16 }
17 ModelId::Gemini3FlashPreview => {
18 "Our most intelligent model built for speed, combining frontier intelligence with superior search and grounding"
19 }
20 ModelId::GPT5 => "Latest most capable OpenAI model with advanced reasoning",
22 ModelId::GPT52 => {
23 "Latest flagship OpenAI model with improved reasoning, xhigh effort, and built-in compaction support"
24 }
25 ModelId::GPT5Mini => "Latest efficient OpenAI model, great for most tasks",
26 ModelId::GPT5Nano => "Latest most cost-effective OpenAI model",
27 ModelId::GPT53Codex => {
28 "GPT-5.3 variant optimized for agentic coding tasks with reasoning effort support (low, medium, high, xhigh)"
29 }
30 ModelId::OpenAIGptOss20b => {
31 "OpenAI's open-source 20B parameter GPT-OSS model using harmony tokenization"
32 }
33 ModelId::OpenAIGptOss120b => {
34 "OpenAI's open-source 120B parameter GPT-OSS model using harmony tokenization"
35 }
36 ModelId::ClaudeOpus46 => {
38 "Next-gen Anthropic flagship with extended and adaptive thinking support"
39 }
40 ModelId::ClaudeSonnet46 => {
41 "Balanced flagship model for coding with extended and adaptive thinking support"
42 }
43 ModelId::ClaudeOpus45 => {
44 "Latest flagship Anthropic model with exceptional reasoning capabilities"
45 }
46 ModelId::ClaudeOpus41 => {
47 "Latest flagship Anthropic model with exceptional reasoning capabilities"
48 }
49 ModelId::ClaudeSonnet45 => "Latest balanced Anthropic model for general tasks",
50 ModelId::ClaudeHaiku45 => {
51 "Latest efficient Anthropic model optimized for low-latency agent workflows"
52 }
53 ModelId::ClaudeSonnet4 => {
54 "Previous balanced Anthropic model maintained for compatibility"
55 }
56 ModelId::DeepSeekChat => {
58 "DeepSeek V3.2 - Fast, efficient chat model for immediate responses"
59 }
60 ModelId::DeepSeekReasoner => {
61 "DeepSeek V3.2 - Thinking mode with integrated tool-use and reasoning capability"
62 }
63 ModelId::ZaiGlm5 => {
65 "Z.ai flagship GLM-5 foundation model engineered for complex systems design and long-horizon agent workflows"
66 }
67 ModelId::OllamaGptOss20b => {
69 "Local GPT-OSS 20B deployment served via Ollama with no external API dependency"
70 }
71 ModelId::OllamaGptOss20bCloud => {
72 "Cloud-hosted GPT-OSS 20B accessed through Ollama Cloud for efficient reasoning tasks"
73 }
74 ModelId::OllamaGptOss120bCloud => {
75 "Cloud-hosted GPT-OSS 120B accessed through Ollama Cloud for larger reasoning tasks"
76 }
77 ModelId::OllamaQwen317b => {
78 "Qwen3 1.7B served locally through Ollama without external API requirements"
79 }
80 ModelId::OllamaQwen3CoderNext => {
81 "Qwen3-Coder-Next served via Ollama Cloud with 256K context, strong coding/tool-use performance, and non-thinking mode responses"
82 }
83 ModelId::OllamaDeepseekV32Cloud => {
84 "DeepSeek V3.2 cloud deployment via Ollama with enhanced reasoning and instruction following"
85 }
86 ModelId::OllamaQwen3Next80bCloud => {
87 "Qwen3 Next generation 80B model via Ollama Cloud with improved reasoning and long context"
88 }
89 ModelId::OllamaMistralLarge3675bCloud => {
90 "Mistral Large 3 675B reasoning model via Ollama Cloud for complex problem-solving"
91 }
92 ModelId::OllamaGlm5Cloud => "Cloud-hosted GLM-5 model served via Ollama Cloud",
93 ModelId::OllamaMinimaxM25Cloud => {
94 "Exceptional multilingual capabilities to elevate code engineering"
95 }
96 ModelId::OllamaGemini3FlashPreviewCloud => {
97 "Gemini 3 Flash offers frontier intelligence built for speed at a fraction of the cost."
98 }
99 ModelId::OllamaDevstral2123bCloud => {
100 "Mistral Devstral 2 123B cloud deployment via Ollama for advanced coding workflows"
101 }
102 ModelId::OllamaNemotron3Nano30bCloud => {
103 "NVIDIA Nemotron-3-Nano 30B brings efficient excellence to code"
104 }
105 ModelId::OllamaQwen3Coder480bCloud => {
106 "Cloud-hosted Qwen3 Coder 480B model accessed through Ollama Cloud for coding tasks"
107 }
108 ModelId::OllamaMinimaxM2Cloud => {
109 "Cloud-hosted MiniMax-M2 model accessed through Ollama Cloud for reasoning tasks"
110 }
111 ModelId::MinimaxM25 => {
112 "Latest MiniMax-M2.5 model with further improvements in reasoning and coding"
113 }
114 ModelId::MinimaxM2 => {
115 "MiniMax-M2 via Anthropic-compatible API with reasoning and tool use"
116 }
117 ModelId::MoonshotMinimaxM25 => "MiniMax-M2.5 served via Moonshot API",
118 ModelId::MoonshotQwen3CoderNext => "Qwen3 Coder Next model served via Moonshot API",
119 ModelId::HuggingFaceDeepseekV32 => {
120 "DeepSeek-V3.2 via Hugging Face router for advanced reasoning"
121 }
122 ModelId::HuggingFaceOpenAIGptOss20b => "OpenAI GPT-OSS 20B via Hugging Face router",
123 ModelId::HuggingFaceOpenAIGptOss120b => "OpenAI GPT-OSS 120B via Hugging Face router",
124 ModelId::HuggingFaceMinimaxM25Novita => {
125 "MiniMax-M2.5 model via Novita inference provider on HuggingFace router. Enhanced reasoning capabilities."
126 }
127 ModelId::HuggingFaceDeepseekV32Novita => {
128 "DeepSeek-V3.2 via Novita inference provider on HuggingFace router."
129 }
130 ModelId::HuggingFaceXiaomiMimoV2FlashNovita => {
131 "Xiaomi MiMo-V2-Flash via Novita on HuggingFace router."
132 }
133 ModelId::HuggingFaceGlm5Novita => {
134 "Z.ai GLM-5 flagship model via Novita inference provider on HuggingFace router."
135 }
136 ModelId::HuggingFaceQwen3CoderNextNovita => {
137 "Qwen3-Coder-Next via Novita inference provider on HuggingFace router. Coding-optimized model with reasoning capabilities."
138 }
139 ModelId::HuggingFaceQwen35397BA17BTogether => {
140 "Qwen3.5-397B-A17B via Together inference provider on HuggingFace router. Vision-language model with linear attention and sparse MoE, 1M context window."
141 }
142 ModelId::OpenRouterMinimaxM25 => "MiniMax-M2.5 flagship model via OpenRouter",
143 ModelId::OpenRouterQwen3CoderNext => {
144 "Next-generation Qwen3 coding model optimized for agentic workflows via OpenRouter"
145 }
146 _ => unreachable!(),
147 }
148 }
149}