vtcode_config/models/model_id.rs
1use serde::{Deserialize, Serialize};
2
3mod as_str;
4mod capabilities;
5mod collection;
6mod defaults;
7mod description;
8mod display;
9mod format;
10mod openrouter;
11mod parse;
12mod provider;
13
14pub use capabilities::{
15 ModelCatalogEntry, catalog_provider_keys, model_catalog_entry, supported_models_for_provider,
16};
17
18/// Centralized enum for all supported model identifiers
19#[cfg_attr(feature = "schema", derive(schemars::JsonSchema))]
20#[derive(Clone, Copy, Debug, Default, PartialEq, Eq, Hash, Serialize, Deserialize)]
21pub enum ModelId {
22 // Gemini models
23 /// Gemini 3.1 Pro Preview - Latest Gemini 3.1 Pro flagship
24 Gemini31ProPreview,
25 /// Gemini 3.1 Pro Preview Custom Tools - Optimized for custom tools & bash
26 Gemini31ProPreviewCustomTools,
27 /// Gemini 3.1 Flash Lite Preview - Most cost-efficient model, offering fastest performance for high-frequency, lightweight tasks
28 Gemini31FlashLitePreview,
29 /// Gemini 3 Flash Preview - Our most intelligent model built for speed, combining frontier intelligence with superior search and grounding
30 #[default]
31 Gemini3FlashPreview,
32
33 // OpenAI models
34 /// GPT-5 - Latest most capable OpenAI model (2025-08-07)
35 GPT5,
36 /// GPT-5.2 - Latest flagship general-purpose OpenAI model (2025-12-11)
37 GPT52,
38 /// GPT-5.2 Codex - Code-focused GPT-5.2 variant optimized for agentic coding with xhigh reasoning support
39 GPT52Codex,
40 /// GPT-5.4 - Mainline frontier GPT model for general-purpose and coding work
41 GPT54,
42 /// GPT-5.4 Pro - Higher-compute GPT-5.4 variant for difficult problems
43 GPT54Pro,
44 /// GPT-5.4 Nano - Lightweight GPT-5.4 variant optimized for speed and cost-efficiency
45 GPT54Nano,
46 /// GPT-5.4 Mini - Compact GPT-5.4 variant for cost-effective tasks with reduced reasoning overhead
47 GPT54Mini,
48 /// GPT-5.3 Codex - Code-focused GPT-5.3 variant optimized for agentic coding with reasoning effort support (low, medium, high, xhigh)
49 GPT53Codex,
50 /// GPT-5.1 Codex - Code-focused GPT-5.1 variant optimized for agentic coding
51 GPT51Codex,
52 /// GPT-5.1 Codex Max - Higher-compute GPT-5.1 Codex variant for longer-running engineering tasks
53 GPT51CodexMax,
54 /// GPT-5 Mini - Latest efficient OpenAI model (2025-08-07)
55 GPT5Mini,
56 /// GPT-5 Nano - Latest most cost-effective OpenAI model (2025-08-07)
57 GPT5Nano,
58 /// GPT-5 Codex - Code-focused GPT-5 variant optimized for agentic coding
59 GPT5Codex,
60 /// GPT-OSS 20B - OpenAI's open-source 20B parameter model using harmony
61 OpenAIGptOss20b,
62 /// GPT-OSS 120B - OpenAI's open-source 120B parameter model using harmony
63 OpenAIGptOss120b,
64
65 // Anthropic models
66 /// Claude Opus 4.6 - Next-gen flagship Anthropic model with extended thinking
67 ClaudeOpus46,
68 /// Claude Sonnet 4.6 - Balanced flagship Anthropic model with extended thinking
69 ClaudeSonnet46,
70 /// Claude Haiku 4.5 - Latest efficient Anthropic model (2025-10-15)
71 ClaudeHaiku45,
72 /// GitHub Copilot auto model selection
73 CopilotAuto,
74 /// GitHub Copilot GPT-5.2 Codex
75 CopilotGPT52Codex,
76 /// GitHub Copilot GPT-5.1 Codex Max
77 CopilotGPT51CodexMax,
78 /// GitHub Copilot GPT-5.4
79 CopilotGPT54,
80 /// GitHub Copilot GPT-5.4 Mini
81 CopilotGPT54Mini,
82 /// GitHub Copilot Claude Sonnet 4.6
83 CopilotClaudeSonnet46,
84
85 // DeepSeek models
86 /// DeepSeek V3.2 Chat - Fast non-thinking mode
87 DeepSeekChat,
88 /// DeepSeek V3.2 Reasoner - Thinking mode with structured reasoning output
89 DeepSeekReasoner,
90 // Hugging Face models
91 /// DeepSeek V3.2 via Hugging Face router
92 HuggingFaceDeepseekV32,
93 /// OpenAI GPT-OSS 20B via Hugging Face router
94 HuggingFaceOpenAIGptOss20b,
95 /// OpenAI GPT-OSS 120B via Hugging Face router
96 HuggingFaceOpenAIGptOss120b,
97 /// DeepSeek V3.2 via Novita on Hugging Face router
98 HuggingFaceDeepseekV32Novita,
99 /// Xiaomi MiMo-V2-Flash via Novita on Hugging Face router
100 HuggingFaceXiaomiMimoV2FlashNovita,
101 /// MiniMax M2.5 via Novita on Hugging Face router
102 HuggingFaceMinimaxM25Novita,
103 /// Z.AI GLM-5 via Novita on Hugging Face router
104 HuggingFaceGlm5Novita,
105 /// Qwen3-Coder-Next via Novita inference provider on Hugging Face router
106 HuggingFaceQwen3CoderNextNovita,
107 /// Qwen3.5-397B-A17B via Together inference provider on Hugging Face router
108 HuggingFaceQwen35397BA17BTogether,
109 /// Step 3.5 Flash via Hugging Face router
110 HuggingFaceStep35Flash,
111
112 /// GLM-5 - Flagship Z.ai foundation model for complex systems
113 ZaiGlm5,
114
115 // Moonshot models
116 /// Kimi K2.5 - Moonshot.ai's flagship reasoning model
117 MoonshotKimiK25,
118
119 // Ollama models
120 /// GPT-OSS 20B - Open-weight GPT-OSS 20B model served via Ollama locally
121 OllamaGptOss20b,
122 /// GPT-OSS 20B Cloud - Cloud-hosted GPT-OSS 20B served via Ollama Cloud
123 OllamaGptOss20bCloud,
124 /// GPT-OSS 120B Cloud - Cloud-hosted GPT-OSS 120B served via Ollama Cloud
125 OllamaGptOss120bCloud,
126 /// Qwen3 1.7B - Qwen3 1.7B model served via Ollama
127 OllamaQwen317b,
128 /// Qwen3 Coder Next - Coding-optimized Qwen3 Next model served via Ollama locally
129 OllamaQwen3CoderNext,
130 /// DeepSeek V3.2 Cloud - DeepSeek V3.2 reasoning deployment via Ollama Cloud
131 OllamaDeepseekV32Cloud,
132 /// Qwen3 Next 80B Cloud - Next-generation Qwen3 80B via Ollama Cloud
133 OllamaQwen3Next80bCloud,
134 /// MiniMax-M2 Cloud - Cloud-hosted MiniMax-M2 model served via Ollama Cloud
135 OllamaMinimaxM2Cloud,
136 /// GLM-5 Cloud - Cloud-hosted GLM-5 model served via Ollama Cloud
137 OllamaGlm5Cloud,
138 /// MiniMax-M2.5 Cloud - Cloud-hosted MiniMax-M2.5 model served via Ollama Cloud
139 OllamaMinimaxM25Cloud,
140 /// Gemini 3 Flash Preview Cloud - Google Gemini 3 Flash Preview via Ollama Cloud
141 OllamaGemini3FlashPreviewCloud,
142 /// Nemotron 3 Super Cloud - NVIDIA Nemotron 3 Super 120B via Ollama Cloud
143 OllamaNemotron3SuperCloud,
144
145 // MiniMax models
146 /// MiniMax-M2.7 - Recursive self-improvement flagship with 204.8K context
147 MinimaxM27,
148 /// MiniMax-M2.5 - Latest MiniMax model with further improvements in reasoning and coding
149 MinimaxM25,
150
151 // OpenRouter models
152 /// Qwen3 32B - Dense 32B Qwen3 deployment
153 OpenRouterQwen332b,
154 /// Qwen3 30B A3B - Active-parameter 30B Qwen3 model
155 OpenRouterQwen330bA3b,
156 /// Qwen3 30B A3B Instruct 2507 - Instruction-tuned Qwen3 30B A3B
157 OpenRouterQwen330bA3bInstruct2507,
158 /// Qwen3 30B A3B Thinking 2507 - Deliberative Qwen3 30B A3B release
159 OpenRouterQwen330bA3bThinking2507,
160 /// Qwen3 14B - Lightweight Qwen3 14B model
161 OpenRouterQwen314b,
162 /// Qwen3 8B - Compact Qwen3 8B deployment
163 OpenRouterQwen38b,
164 /// Qwen3 Next 80B A3B Instruct - Next-generation Qwen3 instruction model
165 OpenRouterQwen3Next80bA3bInstruct,
166 /// Qwen3 Next 80B A3B Thinking - Next-generation Qwen3 reasoning release
167 OpenRouterQwen3Next80bA3bThinking,
168 /// Qwen3.5-397B-A17B - Native vision-language model with linear attention and sparse MoE, 1M context window
169 OpenRouterQwen35Plus0215,
170 /// Qwen3 Coder - Qwen3-based coding model tuned for IDE workflows
171 OpenRouterQwen3Coder,
172 /// Qwen3 Coder Plus - Premium Qwen3 coding model with long context
173 OpenRouterQwen3CoderPlus,
174 /// Qwen3 Coder Flash - Latency optimised Qwen3 coding model
175 OpenRouterQwen3CoderFlash,
176 /// Qwen3 Coder 30B A3B Instruct - Large Mixture-of-Experts coding deployment
177 OpenRouterQwen3Coder30bA3bInstruct,
178 /// Qwen3 Coder Next - Next-generation Qwen3 coding model with enhanced reasoning
179 OpenRouterQwen3CoderNext,
180 /// DeepSeek V3.2 Chat - Official chat model via OpenRouter
181 OpenRouterDeepseekChat,
182 /// DeepSeek V3.2 - Standard model with thinking support via OpenRouter
183 OpenRouterDeepSeekV32,
184 /// DeepSeek V3.2 Reasoner - Thinking mode via OpenRouter
185 OpenRouterDeepseekReasoner,
186 /// DeepSeek V3.2 Speciale - Enhanced reasoning model (no tool-use)
187 OpenRouterDeepSeekV32Speciale,
188 /// DeepSeek V3.2 Exp - Experimental DeepSeek V3.2 listing
189 OpenRouterDeepSeekV32Exp,
190 /// DeepSeek Chat v3.1 - Advanced DeepSeek model via OpenRouter
191 OpenRouterDeepSeekChatV31,
192 /// DeepSeek R1 - DeepSeek R1 reasoning model with chain-of-thought
193 OpenRouterDeepSeekR1,
194 /// OpenAI gpt-oss-120b - Open-weight 120B reasoning model via OpenRouter
195 OpenRouterOpenAIGptOss120b,
196 /// OpenAI gpt-oss-120b:free - Open-weight 120B reasoning model free tier via OpenRouter
197 OpenRouterOpenAIGptOss120bFree,
198 /// OpenAI gpt-oss-20b - Open-weight 20B deployment via OpenRouter
199 OpenRouterOpenAIGptOss20b,
200 /// OpenAI GPT-5 - OpenAI GPT-5 model accessed through OpenRouter
201 OpenRouterOpenAIGpt5,
202 /// OpenAI GPT-5 Chat - Chat optimised GPT-5 endpoint without tool use
203 OpenRouterOpenAIGpt5Chat,
204
205 /// Gemini 3.1 Pro Preview - Google's latest Gemini 3.1 Pro model via OpenRouter
206 OpenRouterGoogleGemini31ProPreview,
207
208 /// Claude Sonnet 4.5 - Anthropic Claude Sonnet 4.5 listing
209 OpenRouterAnthropicClaudeSonnet45,
210 /// Claude Sonnet 4.6 - Anthropic Claude Sonnet 4.6 listing
211 OpenRouterAnthropicClaudeSonnet46,
212 /// Claude Haiku 4.5 - Anthropic Claude Haiku 4.5 listing
213 OpenRouterAnthropicClaudeHaiku45,
214 /// Mistral Large 3 2512 - Mistral Large 3 2512 model via OpenRouter
215 OpenRouterMistralaiMistralLarge2512,
216 /// DeepSeek V3.1 Nex N1 - Nex AGI DeepSeek V3.1 Nex N1 model via OpenRouter
217 OpenRouterNexAgiDeepseekV31NexN1,
218 /// Step 3.5 Flash (free) - StepFun's most capable open-source reasoning model via OpenRouter
219 OpenRouterStepfunStep35FlashFree,
220 /// Nemotron 3 Super (free) - NVIDIA's 120B hybrid MoE model via OpenRouter
221 OpenRouterNvidiaNemotron3Super120bA12bFree,
222 /// GLM-5 - Z.AI GLM-5 flagship foundation model via OpenRouter
223 OpenRouterZaiGlm5,
224 /// MiniMax-M2.5 - MiniMax flagship model via OpenRouter
225 OpenRouterMinimaxM25,
226}