diff --git a/packages/ai/src/models.generated.ts b/packages/ai/src/models.generated.ts index f8062ad5..3d8da58e 100644 --- a/packages/ai/src/models.generated.ts +++ b/packages/ai/src/models.generated.ts @@ -6069,40 +6069,6 @@ export const MODELS = { contextWindow: 400000, maxTokens: 128000, } satisfies Model<"openai-responses">, - "kimi-k2": { - id: "kimi-k2", - name: "Kimi K2", - api: "openai-completions", - provider: "opencode", - baseUrl: "https://opencode.ai/zen/v1", - reasoning: false, - input: ["text"], - cost: { - input: 0.4, - output: 2.5, - cacheRead: 0.4, - cacheWrite: 0, - }, - contextWindow: 262144, - maxTokens: 262144, - } satisfies Model<"openai-completions">, - "kimi-k2-thinking": { - id: "kimi-k2-thinking", - name: "Kimi K2 Thinking", - api: "openai-completions", - provider: "opencode", - baseUrl: "https://opencode.ai/zen/v1", - reasoning: true, - input: ["text"], - cost: { - input: 0.4, - output: 2.5, - cacheRead: 0.4, - cacheWrite: 0, - }, - contextWindow: 262144, - maxTokens: 262144, - } satisfies Model<"openai-completions">, "kimi-k2.5": { id: "kimi-k2.5", name: "Kimi K2.5", @@ -6777,13 +6743,13 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.19, - output: 0.87, - cacheRead: 0.095, + input: 0.19999999999999998, + output: 0.77, + cacheRead: 0.135, cacheWrite: 0, }, contextWindow: 163840, - maxTokens: 65536, + maxTokens: 4096, } satisfies Model<"openai-completions">, "deepseek/deepseek-chat-v3.1": { id: "deepseek/deepseek-chat-v3.1", @@ -6828,9 +6794,9 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.39999999999999997, - output: 1.75, - cacheRead: 0.19999999999999998, + input: 0.44999999999999996, + output: 2.1500000000000004, + cacheRead: 0.22499999999999998, cacheWrite: 0, }, contextWindow: 163840, @@ -6885,7 +6851,7 @@ export const MODELS = { cacheWrite: 0, }, contextWindow: 163840, - maxTokens: 163840, + maxTokens: 65536, } satisfies Model<"openai-completions">, "deepseek/deepseek-v3.2-exp": { id: "deepseek/deepseek-v3.2-exp", @@ -6987,7 +6953,7 @@ export const MODELS = { cacheWrite: 0.08333333333333334, }, contextWindow: 1048576, - maxTokens: 65535, + maxTokens: 65536, } satisfies Model<"openai-completions">, "google/gemini-2.5-pro": { id: "google/gemini-2.5-pro", @@ -7055,7 +7021,7 @@ export const MODELS = { cacheWrite: 0.08333333333333334, }, contextWindow: 1048576, - maxTokens: 65535, + maxTokens: 65536, } satisfies Model<"openai-completions">, "google/gemini-3-pro-preview": { id: "google/gemini-3-pro-preview", @@ -9617,12 +9583,12 @@ export const MODELS = { input: ["text", "image"], cost: { input: 0.25, - output: 2, + output: 1, cacheRead: 0, cacheWrite: 0, }, contextWindow: 262144, - maxTokens: 65536, + maxTokens: 262144, } satisfies Model<"openai-completions">, "qwen/qwen3.5-397b-a17b": { id: "qwen/qwen3.5-397b-a17b", @@ -10177,13 +10143,13 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.95, - output: 2.5500000000000003, - cacheRead: 0.19999999999999998, + input: 0.7999999999999999, + output: 2.56, + cacheRead: 0.16, cacheWrite: 0, }, - contextWindow: 204800, - maxTokens: 131072, + contextWindow: 202752, + maxTokens: 4096, } satisfies Model<"openai-completions">, }, "vercel-ai-gateway": {