vtcode_config/models/model_id.rs
1use serde::{Deserialize, Serialize};
2
3mod as_str;
4mod capabilities;
5mod collection;
6mod defaults;
7mod description;
8mod display;
9mod format;
10mod openrouter;
11mod parse;
12mod provider;
13
14/// Centralized enum for all supported model identifiers
15#[cfg_attr(feature = "schema", derive(schemars::JsonSchema))]
16#[derive(Clone, Copy, Debug, Default, PartialEq, Eq, Hash, Serialize, Deserialize)]
17pub enum ModelId {
18 // Gemini models
19 /// Gemini 3.1 Pro Preview - Latest Gemini 3.1 Pro flagship
20 Gemini31ProPreview,
21 /// Gemini 3.1 Pro Preview Custom Tools - Optimized for custom tools & bash
22 Gemini31ProPreviewCustomTools,
23 /// Gemini 3 Flash Preview - Our most intelligent model built for speed, combining frontier intelligence with superior search and grounding
24 #[default]
25 Gemini3FlashPreview,
26
27 // OpenAI models
28 /// GPT-5 - Latest most capable OpenAI model (2025-08-07)
29 GPT5,
30 /// GPT-5.2 - Latest flagship general-purpose OpenAI model (2025-12-11)
31 GPT52,
32 /// GPT-5 Mini - Latest efficient OpenAI model (2025-08-07)
33 GPT5Mini,
34 /// GPT-5 Nano - Latest most cost-effective OpenAI model (2025-08-07)
35 GPT5Nano,
36 /// GPT-5.3 Codex - Code-focused GPT-5.3 variant optimized for agentic coding with reasoning effort support (low, medium, high, xhigh)
37 GPT53Codex,
38 /// GPT-OSS 20B - OpenAI's open-source 20B parameter model using harmony
39 OpenAIGptOss20b,
40 /// GPT-OSS 120B - OpenAI's open-source 120B parameter model using harmony
41 OpenAIGptOss120b,
42
43 // Anthropic models
44 /// Claude Opus 4.6 - Next-gen flagship Anthropic model with extended thinking
45 ClaudeOpus46,
46 /// Claude Sonnet 4.6 - Balanced flagship Anthropic model with extended thinking
47 ClaudeSonnet46,
48 /// Claude Opus 4.5 - Latest flagship Anthropic model with exceptional reasoning (2025-11-01)
49 ClaudeOpus45,
50 /// Claude Opus 4.1 - Previous most capable Anthropic model (2025-08-05)
51 ClaudeOpus41,
52 /// Claude Sonnet 4.5 - Latest balanced Anthropic model (2025-10-15)
53 ClaudeSonnet45,
54 /// Claude Haiku 4.5 - Latest efficient Anthropic model (2025-10-15)
55 ClaudeHaiku45,
56 /// Claude Sonnet 4 - Previous balanced Anthropic model (2025-05-14)
57 ClaudeSonnet4,
58
59 // DeepSeek models
60 /// DeepSeek V3.2 Chat - Fast non-thinking mode
61 DeepSeekChat,
62 /// DeepSeek V3.2 Reasoner - Thinking mode with structured reasoning output
63 DeepSeekReasoner,
64 // Hugging Face models
65 /// DeepSeek V3.2 via Hugging Face router
66 HuggingFaceDeepseekV32,
67 /// OpenAI GPT-OSS 20B via Hugging Face router
68 HuggingFaceOpenAIGptOss20b,
69 /// OpenAI GPT-OSS 120B via Hugging Face router
70 HuggingFaceOpenAIGptOss120b,
71 /// DeepSeek V3.2 via Novita on Hugging Face router
72 HuggingFaceDeepseekV32Novita,
73 /// Xiaomi MiMo-V2-Flash via Novita on Hugging Face router
74 HuggingFaceXiaomiMimoV2FlashNovita,
75 /// MiniMax M2.5 via Novita on Hugging Face router
76 HuggingFaceMinimaxM25Novita,
77 /// Z.AI GLM-5 via Novita on Hugging Face router
78 HuggingFaceGlm5Novita,
79 /// Qwen3-Coder-Next via Novita inference provider on Hugging Face router
80 HuggingFaceQwen3CoderNextNovita,
81 /// Qwen3.5-397B-A17B via Together inference provider on Hugging Face router
82 HuggingFaceQwen35397BA17BTogether,
83
84 /// GLM-5 - Flagship Z.ai foundation model for complex systems
85 ZaiGlm5,
86
87 // Moonshot models
88 /// MiniMax-M2.5 - MiniMax model served via Moonshot API
89 MoonshotMinimaxM25,
90 /// Qwen3-Coder-Next - Qwen3 Coder Next model served via Moonshot API
91 MoonshotQwen3CoderNext,
92
93 // Ollama models
94 /// GPT-OSS 20B - Open-weight GPT-OSS 20B model served via Ollama locally
95 OllamaGptOss20b,
96 /// GPT-OSS 20B Cloud - Cloud-hosted GPT-OSS 20B served via Ollama Cloud
97 OllamaGptOss20bCloud,
98 /// GPT-OSS 120B Cloud - Cloud-hosted GPT-OSS 120B served via Ollama Cloud
99 OllamaGptOss120bCloud,
100 /// Qwen3 1.7B - Qwen3 1.7B model served via Ollama
101 OllamaQwen317b,
102 /// Qwen3 Coder Next - Coding-optimized Qwen3 Next model served via Ollama locally
103 OllamaQwen3CoderNext,
104 /// DeepSeek V3.2 Cloud - DeepSeek V3.2 reasoning deployment via Ollama Cloud
105 OllamaDeepseekV32Cloud,
106 /// Qwen3 Next 80B Cloud - Next-generation Qwen3 80B via Ollama Cloud
107 OllamaQwen3Next80bCloud,
108 /// Mistral Large 3 675B Cloud - Mistral Large 3 reasoning model via Ollama Cloud
109 OllamaMistralLarge3675bCloud,
110 /// Qwen3 Coder 480B Cloud - Cloud-hosted Qwen3 Coder model served via Ollama Cloud
111 OllamaQwen3Coder480bCloud,
112 /// Devstral 2 123B Cloud - Mistral Devstral 2 123B model via Ollama Cloud
113 OllamaDevstral2123bCloud,
114 /// MiniMax-M2 Cloud - Cloud-hosted MiniMax-M2 model served via Ollama Cloud
115 OllamaMinimaxM2Cloud,
116 /// GLM-5 Cloud - Cloud-hosted GLM-5 model served via Ollama Cloud
117 OllamaGlm5Cloud,
118 /// MiniMax-M2.5 Cloud - Cloud-hosted MiniMax-M2.5 model served via Ollama Cloud
119 OllamaMinimaxM25Cloud,
120 /// Gemini 3 Flash Preview Cloud - Google Gemini 3 Flash Preview via Ollama Cloud
121 OllamaGemini3FlashPreviewCloud,
122 /// Nemotron-3-Nano 30B Cloud - NVIDIA Nemotron-3-Nano 30B via Ollama Cloud
123 OllamaNemotron3Nano30bCloud,
124
125 // MiniMax models
126 /// MiniMax-M2.5 - Latest MiniMax model with further improvements in reasoning and coding
127 MinimaxM25,
128 /// MiniMax-M2 - MiniMax reasoning-focused model
129 MinimaxM2,
130
131 // OpenRouter models
132 /// Qwen3 Max - Flagship Qwen3 mixture for general reasoning
133 OpenRouterQwen3Max,
134 /// Qwen3 235B A22B - Mixture-of-experts Qwen3 235B general model
135 OpenRouterQwen3235bA22b,
136 /// Qwen3 235B A22B Instruct 2507 - Instruction-tuned Qwen3 235B A22B
137 OpenRouterQwen3235bA22b2507,
138 /// Qwen3 235B A22B Thinking 2507 - Deliberative Qwen3 235B A22B reasoning release
139 OpenRouterQwen3235bA22bThinking2507,
140 /// Qwen3 32B - Dense 32B Qwen3 deployment
141 OpenRouterQwen332b,
142 /// Qwen3 30B A3B - Active-parameter 30B Qwen3 model
143 OpenRouterQwen330bA3b,
144 /// Qwen3 30B A3B Instruct 2507 - Instruction-tuned Qwen3 30B A3B
145 OpenRouterQwen330bA3bInstruct2507,
146 /// Qwen3 30B A3B Thinking 2507 - Deliberative Qwen3 30B A3B release
147 OpenRouterQwen330bA3bThinking2507,
148 /// Qwen3 14B - Lightweight Qwen3 14B model
149 OpenRouterQwen314b,
150 /// Qwen3 8B - Compact Qwen3 8B deployment
151 OpenRouterQwen38b,
152 /// Qwen3 Next 80B A3B Instruct - Next-generation Qwen3 instruction model
153 OpenRouterQwen3Next80bA3bInstruct,
154 /// Qwen3 Next 80B A3B Thinking - Next-generation Qwen3 reasoning release
155 OpenRouterQwen3Next80bA3bThinking,
156 /// Qwen3.5-397B-A17B - Native vision-language model with linear attention and sparse MoE, 1M context window
157 OpenRouterQwen35Plus0215,
158 /// Qwen3 Coder - Qwen3-based coding model tuned for IDE workflows
159 OpenRouterQwen3Coder,
160 /// Qwen3 Coder Plus - Premium Qwen3 coding model with long context
161 OpenRouterQwen3CoderPlus,
162 /// Qwen3 Coder Flash - Latency optimised Qwen3 coding model
163 OpenRouterQwen3CoderFlash,
164 /// Qwen3 Coder 30B A3B Instruct - Large Mixture-of-Experts coding deployment
165 OpenRouterQwen3Coder30bA3bInstruct,
166 /// Qwen3 Coder Next - Next-generation Qwen3 coding model with enhanced reasoning
167 OpenRouterQwen3CoderNext,
168 /// DeepSeek V3.2 Chat - Official chat model via OpenRouter
169 OpenRouterDeepseekChat,
170 /// DeepSeek V3.2 - Standard model with thinking support via OpenRouter
171 OpenRouterDeepSeekV32,
172 /// DeepSeek V3.2 Reasoner - Thinking mode via OpenRouter
173 OpenRouterDeepseekReasoner,
174 /// DeepSeek V3.2 Speciale - Enhanced reasoning model (no tool-use)
175 OpenRouterDeepSeekV32Speciale,
176 /// DeepSeek V3.2 Exp - Experimental DeepSeek V3.2 listing
177 OpenRouterDeepSeekV32Exp,
178 /// DeepSeek Chat v3.1 - Advanced DeepSeek model via OpenRouter
179 OpenRouterDeepSeekChatV31,
180 /// DeepSeek R1 - DeepSeek R1 reasoning model with chain-of-thought
181 OpenRouterDeepSeekR1,
182 /// OpenAI gpt-oss-120b - Open-weight 120B reasoning model via OpenRouter
183 OpenRouterOpenAIGptOss120b,
184 /// OpenAI gpt-oss-120b:free - Open-weight 120B reasoning model free tier via OpenRouter
185 OpenRouterOpenAIGptOss120bFree,
186 /// OpenAI gpt-oss-20b - Open-weight 20B deployment via OpenRouter
187 OpenRouterOpenAIGptOss20b,
188 /// OpenAI GPT-5 - OpenAI GPT-5 model accessed through OpenRouter
189 OpenRouterOpenAIGpt5,
190 /// OpenAI GPT-5 Chat - Chat optimised GPT-5 endpoint without tool use
191 OpenRouterOpenAIGpt5Chat,
192
193 /// Gemini 3.1 Pro Preview - Google's latest Gemini 3.1 Pro model via OpenRouter
194 OpenRouterGoogleGemini31ProPreview,
195
196 /// Claude Sonnet 4.5 - Anthropic Claude Sonnet 4.5 listing
197 OpenRouterAnthropicClaudeSonnet45,
198 /// Claude Sonnet 4.6 - Anthropic Claude Sonnet 4.6 listing
199 OpenRouterAnthropicClaudeSonnet46,
200 /// Claude Haiku 4.5 - Anthropic Claude Haiku 4.5 listing
201 OpenRouterAnthropicClaudeHaiku45,
202 /// Claude Opus 4.1 - Anthropic Claude Opus 4.1 listing
203 OpenRouterAnthropicClaudeOpus41,
204 /// Amazon Nova 2 Lite - Amazon Nova 2 Lite model via OpenRouter
205 OpenRouterAmazonNova2LiteV1,
206 /// Mistral Large 3 2512 - Mistral Large 3 2512 model via OpenRouter
207 OpenRouterMistralaiMistralLarge2512,
208 /// DeepSeek V3.1 Nex N1 - Nex AGI DeepSeek V3.1 Nex N1 model via OpenRouter
209 OpenRouterNexAgiDeepseekV31NexN1,
210 /// OpenAI o1-pro - OpenAI o1-pro advanced reasoning model via OpenRouter
211 OpenRouterOpenAIO1Pro,
212 /// Step 3.5 Flash (free) - StepFun's most capable open-source reasoning model via OpenRouter
213 OpenRouterStepfunStep35FlashFree,
214 /// GLM-5 - Z.AI GLM-5 flagship foundation model via OpenRouter
215 OpenRouterZaiGlm5,
216 /// MiniMax-M2.5 - MiniMax flagship model via OpenRouter
217 OpenRouterMinimaxM25,
218}