This commit is contained in:
Harivansh Rathi 2026-03-06 19:58:19 -08:00
parent 6bdf0ec058
commit 4e1cdf1102

View file

@ -2913,6 +2913,24 @@ export const MODELS = {
contextWindow: 400000, contextWindow: 400000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"openai-responses">, } satisfies Model<"openai-responses">,
"gpt-5.4": {
id: "gpt-5.4",
name: "GPT-5.4",
api: "openai-responses",
provider: "github-copilot",
baseUrl: "https://api.individual.githubcopilot.com",
headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"},
reasoning: true,
input: ["text", "image"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 400000,
maxTokens: 128000,
} satisfies Model<"openai-responses">,
"grok-code-fast-1": { "grok-code-fast-1": {
id: "grok-code-fast-1", id: "grok-code-fast-1",
name: "Grok Code Fast 1", name: "Grok Code Fast 1",
@ -7598,7 +7616,7 @@ export const MODELS = {
cost: { cost: {
input: 0.27, input: 0.27,
output: 0.95, output: 0.95,
cacheRead: 0.0299999997, cacheRead: 0.0290000007,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 196608, contextWindow: 196608,
@ -9398,13 +9416,13 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0, input: 0.11,
output: 0, output: 0.6,
cacheRead: 0, cacheRead: 0.055,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 131072, contextWindow: 262144,
maxTokens: 4096, maxTokens: 262144,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-30b-a3b": { "qwen/qwen3-30b-a3b": {
id: "qwen/qwen3-30b-a3b", id: "qwen/qwen3-30b-a3b",
@ -10401,9 +10419,9 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.3, input: 0.38,
output: 1.4, output: 1.9800000000000002,
cacheRead: 0.15, cacheRead: 0.19,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 202752, contextWindow: 202752,
@ -11448,6 +11466,23 @@ export const MODELS = {
contextWindow: 204800, contextWindow: 204800,
maxTokens: 131000, maxTokens: 131000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"minimax/minimax-m2.5-highspeed": {
id: "minimax/minimax-m2.5-highspeed",
name: "MiniMax M2.5 High Speed",
api: "anthropic-messages",
provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true,
input: ["text"],
cost: {
input: 0.6,
output: 2.4,
cacheRead: 0.03,
cacheWrite: 0.375,
},
contextWindow: 4096,
maxTokens: 4096,
} satisfies Model<"anthropic-messages">,
"mistral/codestral": { "mistral/codestral": {
id: "mistral/codestral", id: "mistral/codestral",
name: "Mistral Codestral", name: "Mistral Codestral",