chore(ai): update generated models data

This commit is contained in:
Mario Zechner 2026-03-02 19:41:16 +01:00
parent 46bfa7931e
commit 105e296b69

View file

@ -6069,40 +6069,6 @@ export const MODELS = {
contextWindow: 400000,
maxTokens: 128000,
} satisfies Model<"openai-responses">,
"kimi-k2": {
id: "kimi-k2",
name: "Kimi K2",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.4,
output: 2.5,
cacheRead: 0.4,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 262144,
} satisfies Model<"openai-completions">,
"kimi-k2-thinking": {
id: "kimi-k2-thinking",
name: "Kimi K2 Thinking",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0.4,
output: 2.5,
cacheRead: 0.4,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 262144,
} satisfies Model<"openai-completions">,
"kimi-k2.5": {
id: "kimi-k2.5",
name: "Kimi K2.5",
@ -6777,13 +6743,13 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.19,
output: 0.87,
cacheRead: 0.095,
input: 0.19999999999999998,
output: 0.77,
cacheRead: 0.135,
cacheWrite: 0,
},
contextWindow: 163840,
maxTokens: 65536,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"deepseek/deepseek-chat-v3.1": {
id: "deepseek/deepseek-chat-v3.1",
@ -6828,9 +6794,9 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.39999999999999997,
output: 1.75,
cacheRead: 0.19999999999999998,
input: 0.44999999999999996,
output: 2.1500000000000004,
cacheRead: 0.22499999999999998,
cacheWrite: 0,
},
contextWindow: 163840,
@ -6885,7 +6851,7 @@ export const MODELS = {
cacheWrite: 0,
},
contextWindow: 163840,
maxTokens: 163840,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"deepseek/deepseek-v3.2-exp": {
id: "deepseek/deepseek-v3.2-exp",
@ -6987,7 +6953,7 @@ export const MODELS = {
cacheWrite: 0.08333333333333334,
},
contextWindow: 1048576,
maxTokens: 65535,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"google/gemini-2.5-pro": {
id: "google/gemini-2.5-pro",
@ -7055,7 +7021,7 @@ export const MODELS = {
cacheWrite: 0.08333333333333334,
},
contextWindow: 1048576,
maxTokens: 65535,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"google/gemini-3-pro-preview": {
id: "google/gemini-3-pro-preview",
@ -9617,12 +9583,12 @@ export const MODELS = {
input: ["text", "image"],
cost: {
input: 0.25,
output: 2,
output: 1,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 65536,
maxTokens: 262144,
} satisfies Model<"openai-completions">,
"qwen/qwen3.5-397b-a17b": {
id: "qwen/qwen3.5-397b-a17b",
@ -10177,13 +10143,13 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.95,
output: 2.5500000000000003,
cacheRead: 0.19999999999999998,
input: 0.7999999999999999,
output: 2.56,
cacheRead: 0.16,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
contextWindow: 202752,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
},
"vercel-ai-gateway": {