Skip to main content

vtcode_config/models/
model_id.rs

1use serde::{Deserialize, Serialize};
2
3mod as_str;
4mod capabilities;
5mod collection;
6mod defaults;
7mod description;
8mod display;
9mod format;
10mod openrouter;
11mod parse;
12mod provider;
13
14/// Centralized enum for all supported model identifiers
15#[cfg_attr(feature = "schema", derive(schemars::JsonSchema))]
16#[derive(Clone, Copy, Debug, Default, PartialEq, Eq, Hash, Serialize, Deserialize)]
17pub enum ModelId {
18    // Gemini models
19    /// Gemini 3.1 Pro Preview - Latest Gemini 3.1 Pro flagship
20    Gemini31ProPreview,
21    /// Gemini 3.1 Pro Preview Custom Tools - Optimized for custom tools & bash
22    Gemini31ProPreviewCustomTools,
23    /// Gemini 3 Pro Preview - Preview of next-generation Gemini model
24    Gemini3ProPreview,
25    /// Gemini 3 Flash Preview - Our most intelligent model built for speed, combining frontier intelligence with superior search and grounding
26    #[default]
27    Gemini3FlashPreview,
28
29    // OpenAI models
30    /// GPT-5 - Latest most capable OpenAI model (2025-08-07)
31    GPT5,
32    /// GPT-5.2 - Latest flagship general-purpose OpenAI model (2025-12-11)
33    GPT52,
34    /// GPT-5.2 Codex - Code-focused GPT-5.2 variant optimized for agentic coding
35    GPT52Codex,
36    /// GPT-5 Codex - Code-focused GPT-5 variant using the Responses API
37    GPT5Codex,
38    /// GPT-5 Mini - Latest efficient OpenAI model (2025-08-07)
39    GPT5Mini,
40    /// GPT-5 Nano - Latest most cost-effective OpenAI model (2025-08-07)
41    GPT5Nano,
42    /// GPT-5.1 - Enhanced latest most capable OpenAI model with improved reasoning (2025-11-14)
43    GPT51,
44    /// GPT-5.1 Codex - Code-focused GPT-5.1 variant using the Responses API
45    GPT51Codex,
46    /// GPT-5.1 Codex Max - Maximum context code-focused GPT-5.1 variant
47    GPT51CodexMax,
48    /// GPT-5.1 Mini - Enhanced efficient OpenAI model with improved capabilities (2025-11-14)
49    GPT51Mini,
50    /// Codex Mini Latest - Latest Codex model for code generation (2025-05-16)
51    CodexMiniLatest,
52    /// GPT-OSS 20B - OpenAI's open-source 20B parameter model using harmony
53    OpenAIGptOss20b,
54    /// GPT-OSS 120B - OpenAI's open-source 120B parameter model using harmony
55    OpenAIGptOss120b,
56
57    // Anthropic models
58    /// Claude Opus 4.6 - Next-gen flagship Anthropic model with extended thinking
59    ClaudeOpus46,
60    /// Claude Sonnet 4.6 - Balanced flagship Anthropic model with extended thinking
61    ClaudeSonnet46,
62    /// Claude Opus 4.5 - Latest flagship Anthropic model with exceptional reasoning (2025-11-01)
63    ClaudeOpus45,
64    /// Claude Opus 4.1 - Previous most capable Anthropic model (2025-08-05)
65    ClaudeOpus41,
66    /// Claude Sonnet 4.5 - Latest balanced Anthropic model (2025-10-15)
67    ClaudeSonnet45,
68    /// Claude Haiku 4.5 - Latest efficient Anthropic model (2025-10-15)
69    ClaudeHaiku45,
70    /// Claude Sonnet 4 - Previous balanced Anthropic model (2025-05-14)
71    ClaudeSonnet4,
72    /// Claude Opus 4 - Previous flagship model (2025-05-14)
73    ClaudeOpus4,
74    /// Claude Sonnet 3.7 - Latest Claude 3 Sonnet (2025-02-19)
75    ClaudeSonnet37,
76    /// Claude Haiku 3.5 - Latest Claude 3 Haiku (2024-10-22)
77    ClaudeHaiku35,
78
79    // DeepSeek models
80    /// DeepSeek V3.2 Chat - Fast non-thinking mode
81    DeepSeekChat,
82    /// DeepSeek V3.2 Reasoner - Thinking mode with structured reasoning output
83    DeepSeekReasoner,
84    // Hugging Face models
85    /// DeepSeek V3.2 via Hugging Face router
86    HuggingFaceDeepseekV32,
87    /// OpenAI GPT-OSS 20B via Hugging Face router
88    HuggingFaceOpenAIGptOss20b,
89    /// OpenAI GPT-OSS 120B via Hugging Face router
90    HuggingFaceOpenAIGptOss120b,
91    /// DeepSeek V3.2 via Novita on Hugging Face router
92    HuggingFaceDeepseekV32Novita,
93    /// Xiaomi MiMo-V2-Flash via Novita on Hugging Face router
94    HuggingFaceXiaomiMimoV2FlashNovita,
95    /// MiniMax M2.5 via Novita on Hugging Face router
96    HuggingFaceMinimaxM25Novita,
97    /// Z.AI GLM-5 via Novita on Hugging Face router
98    HuggingFaceGlm5Novita,
99    /// Qwen3-Coder-Next via Novita inference provider on Hugging Face router
100    HuggingFaceQwen3CoderNextNovita,
101
102    // xAI models
103    /// Grok-4 - Flagship xAI model with advanced reasoning
104    XaiGrok4,
105    /// Grok-4 Mini - Efficient xAI model variant
106    XaiGrok4Mini,
107    /// Grok-4 Code - Code-focused Grok deployment
108    XaiGrok4Code,
109    /// Grok-4 Code Latest - Latest Grok code model with enhanced reasoning tools
110    XaiGrok4CodeLatest,
111    /// Grok-4 Vision - Multimodal Grok model
112    XaiGrok4Vision,
113
114    /// GLM-5 - Flagship Z.ai foundation model for complex systems
115    ZaiGlm5,
116
117    // Moonshot models
118    /// MiniMax-M2.5 - MiniMax model served via Moonshot API
119    MoonshotMinimaxM25,
120    /// Qwen3-Coder-Next - Qwen3 Coder Next model served via Moonshot API
121    MoonshotQwen3CoderNext,
122
123    // Ollama models
124    /// GPT-OSS 20B - Open-weight GPT-OSS 20B model served via Ollama locally
125    OllamaGptOss20b,
126    /// GPT-OSS 20B Cloud - Cloud-hosted GPT-OSS 20B served via Ollama Cloud
127    OllamaGptOss20bCloud,
128    /// GPT-OSS 120B Cloud - Cloud-hosted GPT-OSS 120B served via Ollama Cloud
129    OllamaGptOss120bCloud,
130    /// Qwen3 1.7B - Qwen3 1.7B model served via Ollama
131    OllamaQwen317b,
132    /// DeepSeek V3.2 Cloud - DeepSeek V3.2 reasoning deployment via Ollama Cloud
133    OllamaDeepseekV32Cloud,
134    /// Qwen3 Next 80B Cloud - Next-generation Qwen3 80B via Ollama Cloud
135    OllamaQwen3Next80bCloud,
136    /// Mistral Large 3 675B Cloud - Mistral Large 3 reasoning model via Ollama Cloud
137    OllamaMistralLarge3675bCloud,
138    /// Qwen3 Coder 480B Cloud - Cloud-hosted Qwen3 Coder model served via Ollama Cloud
139    OllamaQwen3Coder480bCloud,
140    /// Gemini 3 Pro Preview Latest Cloud - Google Gemini 3 Pro Preview via Ollama Cloud
141    OllamaGemini3ProPreviewLatestCloud,
142    /// Devstral 2 123B Cloud - Mistral Devstral 2 123B model via Ollama Cloud
143    OllamaDevstral2123bCloud,
144    /// MiniMax-M2 Cloud - Cloud-hosted MiniMax-M2 model served via Ollama Cloud
145    OllamaMinimaxM2Cloud,
146    /// GLM-5 Cloud - Cloud-hosted GLM-5 model served via Ollama Cloud
147    OllamaGlm5Cloud,
148    /// MiniMax-M2.5 Cloud - Cloud-hosted MiniMax-M2.5 model served via Ollama Cloud
149    OllamaMinimaxM25Cloud,
150    /// Gemini 3 Flash Preview Cloud - Google Gemini 3 Flash Preview via Ollama Cloud
151    OllamaGemini3FlashPreviewCloud,
152    /// Nemotron-3-Nano 30B Cloud - NVIDIA Nemotron-3-Nano 30B via Ollama Cloud
153    OllamaNemotron3Nano30bCloud,
154
155    // MiniMax models
156    /// MiniMax-M2.5 - Latest MiniMax model with further improvements in reasoning and coding
157    MinimaxM25,
158    /// MiniMax-M2 - MiniMax reasoning-focused model
159    MinimaxM2,
160
161    // LM Studio models
162    /// Meta Llama 3 8B Instruct served locally via LM Studio
163    LmStudioMetaLlama38BInstruct,
164    /// Meta Llama 3.1 8B Instruct served locally via LM Studio
165    LmStudioMetaLlama318BInstruct,
166    /// Qwen2.5 7B Instruct served locally via LM Studio
167    LmStudioQwen257BInstruct,
168    /// Gemma 2 2B IT served locally via LM Studio
169    LmStudioGemma22BIt,
170    /// Gemma 2 9B IT served locally via LM Studio
171    LmStudioGemma29BIt,
172    /// Phi-3.1 Mini 4K Instruct served locally via LM Studio
173    LmStudioPhi31Mini4kInstruct,
174
175    // OpenRouter models
176    /// Grok Code Fast 1 - Fast OpenRouter coding model powered by xAI Grok
177    OpenRouterGrokCodeFast1,
178    /// Grok 4 Fast - Reasoning-focused Grok endpoint with transparent traces
179    OpenRouterGrok4Fast,
180    /// Grok 4.1 Fast - Enhanced Grok 4.1 fast inference with improved reasoning
181    OpenRouterGrok41Fast,
182    /// Grok 4 - Flagship Grok 4 endpoint exposed through OpenRouter
183    OpenRouterGrok4,
184    /// Qwen3 Max - Flagship Qwen3 mixture for general reasoning
185    OpenRouterQwen3Max,
186    /// Qwen3 235B A22B - Mixture-of-experts Qwen3 235B general model
187    OpenRouterQwen3235bA22b,
188    /// Qwen3 235B A22B Instruct 2507 - Instruction-tuned Qwen3 235B A22B
189    OpenRouterQwen3235bA22b2507,
190    /// Qwen3 235B A22B Thinking 2507 - Deliberative Qwen3 235B A22B reasoning release
191    OpenRouterQwen3235bA22bThinking2507,
192    /// Qwen3 32B - Dense 32B Qwen3 deployment
193    OpenRouterQwen332b,
194    /// Qwen3 30B A3B - Active-parameter 30B Qwen3 model
195    OpenRouterQwen330bA3b,
196    /// Qwen3 30B A3B Instruct 2507 - Instruction-tuned Qwen3 30B A3B
197    OpenRouterQwen330bA3bInstruct2507,
198    /// Qwen3 30B A3B Thinking 2507 - Deliberative Qwen3 30B A3B release
199    OpenRouterQwen330bA3bThinking2507,
200    /// Qwen3 14B - Lightweight Qwen3 14B model
201    OpenRouterQwen314b,
202    /// Qwen3 8B - Compact Qwen3 8B deployment
203    OpenRouterQwen38b,
204    /// Qwen3 Next 80B A3B Instruct - Next-generation Qwen3 instruction model
205    OpenRouterQwen3Next80bA3bInstruct,
206    /// Qwen3 Next 80B A3B Thinking - Next-generation Qwen3 reasoning release
207    OpenRouterQwen3Next80bA3bThinking,
208    /// Qwen3.5-397B-A17B - Native vision-language model with linear attention and sparse MoE, 1M context window
209    OpenRouterQwen35Plus0215,
210    /// Qwen3 Coder - Qwen3-based coding model tuned for IDE workflows
211    OpenRouterQwen3Coder,
212    /// Qwen3 Coder Plus - Premium Qwen3 coding model with long context
213    OpenRouterQwen3CoderPlus,
214    /// Qwen3 Coder Flash - Latency optimised Qwen3 coding model
215    OpenRouterQwen3CoderFlash,
216    /// Qwen3 Coder 30B A3B Instruct - Large Mixture-of-Experts coding deployment
217    OpenRouterQwen3Coder30bA3bInstruct,
218    /// Qwen3 Coder Next - Next-generation Qwen3 coding model with enhanced reasoning
219    OpenRouterQwen3CoderNext,
220    /// DeepSeek V3.2 Chat - Official chat model via OpenRouter
221    OpenRouterDeepseekChat,
222    /// DeepSeek V3.2 - Standard model with thinking support via OpenRouter
223    OpenRouterDeepSeekV32,
224    /// DeepSeek V3.2 Reasoner - Thinking mode via OpenRouter
225    OpenRouterDeepseekReasoner,
226    /// DeepSeek V3.2 Speciale - Enhanced reasoning model (no tool-use)
227    OpenRouterDeepSeekV32Speciale,
228    /// DeepSeek V3.2 Exp - Experimental DeepSeek V3.2 listing
229    OpenRouterDeepSeekV32Exp,
230    /// DeepSeek Chat v3.1 - Advanced DeepSeek model via OpenRouter
231    OpenRouterDeepSeekChatV31,
232    /// DeepSeek R1 - DeepSeek R1 reasoning model with chain-of-thought
233    OpenRouterDeepSeekR1,
234    /// OpenAI gpt-oss-120b - Open-weight 120B reasoning model via OpenRouter
235    OpenRouterOpenAIGptOss120b,
236    /// OpenAI gpt-oss-120b:free - Open-weight 120B reasoning model free tier via OpenRouter
237    OpenRouterOpenAIGptOss120bFree,
238    /// OpenAI gpt-oss-20b - Open-weight 20B deployment via OpenRouter
239    OpenRouterOpenAIGptOss20b,
240    /// OpenAI GPT-5 - OpenAI GPT-5 model accessed through OpenRouter
241    OpenRouterOpenAIGpt5,
242    /// OpenAI GPT-5 Codex - OpenRouter listing for GPT-5 Codex
243    OpenRouterOpenAIGpt5Codex,
244    /// OpenAI GPT-5 Chat - Chat optimised GPT-5 endpoint without tool use
245    OpenRouterOpenAIGpt5Chat,
246
247    /// Gemini 3.1 Pro Preview - Google's latest Gemini 3.1 Pro model via OpenRouter
248    OpenRouterGoogleGemini31ProPreview,
249
250    /// Claude Sonnet 4.5 - Anthropic Claude Sonnet 4.5 listing
251    OpenRouterAnthropicClaudeSonnet45,
252    /// Claude Sonnet 4.6 - Anthropic Claude Sonnet 4.6 listing
253    OpenRouterAnthropicClaudeSonnet46,
254    /// Claude Haiku 4.5 - Anthropic Claude Haiku 4.5 listing
255    OpenRouterAnthropicClaudeHaiku45,
256    /// Claude Opus 4.1 - Anthropic Claude Opus 4.1 listing
257    OpenRouterAnthropicClaudeOpus41,
258    /// Amazon Nova 2 Lite - Amazon Nova 2 Lite model via OpenRouter
259    OpenRouterAmazonNova2LiteV1,
260    /// Mistral Large 3 2512 - Mistral Large 3 2512 model via OpenRouter
261    OpenRouterMistralaiMistralLarge2512,
262    /// DeepSeek V3.1 Nex N1 - Nex AGI DeepSeek V3.1 Nex N1 model via OpenRouter
263    OpenRouterNexAgiDeepseekV31NexN1,
264    /// OpenAI GPT-5.1 - OpenAI GPT-5.1 model accessed through OpenRouter
265    OpenRouterOpenAIGpt51,
266    /// OpenAI GPT-5.1-Codex - OpenRouter listing for GPT-5.1 Codex
267    OpenRouterOpenAIGpt51Codex,
268    /// OpenAI GPT-5.1-Codex-Max - OpenRouter listing for GPT-5.1 Codex Max
269    OpenRouterOpenAIGpt51CodexMax,
270    /// OpenAI GPT-5.1-Codex-Mini - OpenRouter listing for GPT-5.1 Codex Mini
271    OpenRouterOpenAIGpt51CodexMini,
272    /// OpenAI GPT-5.1 Chat - Chat optimised GPT-5.1 endpoint without tool use
273    OpenRouterOpenAIGpt51Chat,
274    /// OpenAI GPT-5.2 - OpenAI GPT-5.2 model accessed through OpenRouter
275    OpenRouterOpenAIGpt52,
276    /// OpenAI GPT-5.2 Chat - Chat optimised GPT-5.2 endpoint without tool use
277    OpenRouterOpenAIGpt52Chat,
278    /// OpenAI GPT-5.2-Codex - OpenRouter listing for GPT-5.2 Codex
279    OpenRouterOpenAIGpt52Codex,
280    /// OpenAI GPT-5.2 Pro - Professional tier GPT-5.2 model accessed through OpenRouter
281    OpenRouterOpenAIGpt52Pro,
282    /// OpenAI o1-pro - OpenAI o1-pro advanced reasoning model via OpenRouter
283    OpenRouterOpenAIO1Pro,
284    /// Step 3.5 Flash (free) - StepFun's most capable open-source reasoning model via OpenRouter
285    OpenRouterStepfunStep35FlashFree,
286    /// GLM-5 - Z.AI GLM-5 flagship foundation model via OpenRouter
287    OpenRouterZaiGlm5,
288    /// MoonshotAI: Kimi K2 0905 - MoonshotAI Kimi K2 0905 MoE release optimised for coding agents
289    OpenRouterMoonshotaiKimiK20905,
290    /// MoonshotAI: Kimi K2 Thinking - MoonshotAI reasoning-tier Kimi K2 release optimized for long-horizon agents
291    OpenRouterMoonshotaiKimiK2Thinking,
292    /// MoonshotAI: Kimi K2.5 - MoonshotAI Kimi K2.5 multimodal model with long-context and reasoning capabilities via OpenRouter
293    OpenRouterMoonshotaiKimiK25,
294    /// MiniMax-M2.5 - MiniMax flagship model via OpenRouter
295    OpenRouterMinimaxM25,
296}