Add [Unreleased] section for next cycle

This commit is contained in:
Mario Zechner 2026-02-12 22:19:30 +01:00
parent 2668a1125d
commit d0543d3c18
7 changed files with 381 additions and 231 deletions

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
## [0.52.9] - 2026-02-08 ## [0.52.9] - 2026-02-08

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
### Added ### Added

View file

@ -804,6 +804,23 @@ export const MODELS = {
contextWindow: 204608, contextWindow: 204608,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"bedrock-converse-stream">, } satisfies Model<"bedrock-converse-stream">,
"minimax.minimax-m2.1": {
id: "minimax.minimax-m2.1",
name: "MiniMax M2.1",
api: "bedrock-converse-stream",
provider: "amazon-bedrock",
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
reasoning: true,
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"bedrock-converse-stream">,
"mistral.ministral-3-14b-instruct": { "mistral.ministral-3-14b-instruct": {
id: "mistral.ministral-3-14b-instruct", id: "mistral.ministral-3-14b-instruct",
name: "Ministral 14B 3.0", name: "Ministral 14B 3.0",
@ -906,6 +923,23 @@ export const MODELS = {
contextWindow: 256000, contextWindow: 256000,
maxTokens: 256000, maxTokens: 256000,
} satisfies Model<"bedrock-converse-stream">, } satisfies Model<"bedrock-converse-stream">,
"moonshotai.kimi-k2.5": {
id: "moonshotai.kimi-k2.5",
name: "Kimi K2.5",
api: "bedrock-converse-stream",
provider: "amazon-bedrock",
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.6,
output: 3,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 256000,
maxTokens: 256000,
} satisfies Model<"bedrock-converse-stream">,
"nvidia.nemotron-nano-12b-v2": { "nvidia.nemotron-nano-12b-v2": {
id: "nvidia.nemotron-nano-12b-v2", id: "nvidia.nemotron-nano-12b-v2",
name: "NVIDIA Nemotron Nano 12B v2 VL BF16", name: "NVIDIA Nemotron Nano 12B v2 VL BF16",
@ -1229,6 +1263,40 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 64000, maxTokens: 64000,
} satisfies Model<"bedrock-converse-stream">, } satisfies Model<"bedrock-converse-stream">,
"zai.glm-4.7": {
id: "zai.glm-4.7",
name: "GLM-4.7",
api: "bedrock-converse-stream",
provider: "amazon-bedrock",
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
reasoning: true,
input: ["text"],
cost: {
input: 0.6,
output: 2.2,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"bedrock-converse-stream">,
"zai.glm-4.7-flash": {
id: "zai.glm-4.7-flash",
name: "GLM-4.7-Flash",
api: "bedrock-converse-stream",
provider: "amazon-bedrock",
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
reasoning: true,
input: ["text"],
cost: {
input: 0.07,
output: 0.4,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 200000,
maxTokens: 131072,
} satisfies Model<"bedrock-converse-stream">,
}, },
"anthropic": { "anthropic": {
"claude-3-5-haiku-20241022": { "claude-3-5-haiku-20241022": {
@ -2066,6 +2134,23 @@ export const MODELS = {
contextWindow: 400000, contextWindow: 400000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"azure-openai-responses">, } satisfies Model<"azure-openai-responses">,
"gpt-5.3-codex-spark": {
id: "gpt-5.3-codex-spark",
name: "GPT-5.3 Codex Spark",
api: "azure-openai-responses",
provider: "azure-openai-responses",
baseUrl: "",
reasoning: true,
input: ["text", "image"],
cost: {
input: 1.75,
output: 14,
cacheRead: 0.175,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 32000,
} satisfies Model<"azure-openai-responses">,
"o1": { "o1": {
id: "o1", id: "o1",
name: "o1", name: "o1",
@ -2221,6 +2306,23 @@ export const MODELS = {
contextWindow: 131072, contextWindow: 131072,
maxTokens: 32768, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"llama3.1-8b": {
id: "llama3.1-8b",
name: "Llama 3.1 8B",
api: "openai-completions",
provider: "cerebras",
baseUrl: "https://api.cerebras.ai/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.1,
output: 0.1,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 32000,
maxTokens: 8000,
} satisfies Model<"openai-completions">,
"qwen-3-235b-a22b-instruct-2507": { "qwen-3-235b-a22b-instruct-2507": {
id: "qwen-3-235b-a22b-instruct-2507", id: "qwen-3-235b-a22b-instruct-2507",
name: "Qwen 3 235B Instruct", name: "Qwen 3 235B Instruct",
@ -2273,7 +2375,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 128000,
maxTokens: 16000, maxTokens: 32000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"claude-opus-4.5": { "claude-opus-4.5": {
id: "claude-opus-4.5", id: "claude-opus-4.5",
@ -2291,7 +2393,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 128000,
maxTokens: 16000, maxTokens: 32000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"claude-opus-4.6": { "claude-opus-4.6": {
id: "claude-opus-4.6", id: "claude-opus-4.6",
@ -2345,7 +2447,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 128000,
maxTokens: 16000, maxTokens: 32000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"gemini-2.5-pro": { "gemini-2.5-pro": {
id: "gemini-2.5-pro", id: "gemini-2.5-pro",
@ -2420,7 +2522,7 @@ export const MODELS = {
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 64000,
maxTokens: 16384, maxTokens: 16384,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"gpt-4o": { "gpt-4o": {
@ -2494,7 +2596,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 128000,
maxTokens: 128000, maxTokens: 64000,
} satisfies Model<"openai-responses">, } satisfies Model<"openai-responses">,
"gpt-5.1-codex": { "gpt-5.1-codex": {
id: "gpt-5.1-codex", id: "gpt-5.1-codex",
@ -2548,7 +2650,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 128000,
maxTokens: 100000, maxTokens: 128000,
} satisfies Model<"openai-responses">, } satisfies Model<"openai-responses">,
"gpt-5.2": { "gpt-5.2": {
id: "gpt-5.2", id: "gpt-5.2",
@ -3872,6 +3974,24 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"zai-org/GLM-5": {
id: "zai-org/GLM-5",
name: "GLM-5",
api: "openai-completions",
provider: "huggingface",
baseUrl: "https://router.huggingface.co/v1",
compat: {"supportsDeveloperRole":false},
reasoning: true,
input: ["text"],
cost: {
input: 1,
output: 3.2,
cacheRead: 0.2,
cacheWrite: 0,
},
contextWindow: 202752,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
}, },
"kimi-coding": { "kimi-coding": {
"k2p5": { "k2p5": {
@ -4868,6 +4988,23 @@ export const MODELS = {
contextWindow: 400000, contextWindow: 400000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"openai-responses">, } satisfies Model<"openai-responses">,
"gpt-5.3-codex-spark": {
id: "gpt-5.3-codex-spark",
name: "GPT-5.3 Codex Spark",
api: "openai-responses",
provider: "openai",
baseUrl: "https://api.openai.com/v1",
reasoning: true,
input: ["text", "image"],
cost: {
input: 1.75,
output: 14,
cacheRead: 0.175,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 32000,
} satisfies Model<"openai-responses">,
"o1": { "o1": {
id: "o1", id: "o1",
name: "o1", name: "o1",
@ -5108,6 +5245,23 @@ export const MODELS = {
contextWindow: 272000, contextWindow: 272000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"openai-codex-responses">, } satisfies Model<"openai-codex-responses">,
"gpt-5.3-codex-spark": {
id: "gpt-5.3-codex-spark",
name: "GPT-5.3 Codex Spark",
api: "openai-codex-responses",
provider: "openai-codex",
baseUrl: "https://chatgpt.com/backend-api",
reasoning: true,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 128000,
} satisfies Model<"openai-codex-responses">,
}, },
"opencode": { "opencode": {
"big-pickle": { "big-pickle": {
@ -5314,23 +5468,6 @@ export const MODELS = {
contextWindow: 204800, contextWindow: 204800,
maxTokens: 131072, maxTokens: 131072,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"glm-4.7-free": {
id: "glm-4.7-free",
name: "GLM-4.7 Free",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"gpt-5": { "gpt-5": {
id: "gpt-5", id: "gpt-5",
name: "GPT-5", name: "GPT-5",
@ -5569,12 +5706,12 @@ export const MODELS = {
contextWindow: 204800, contextWindow: 204800,
maxTokens: 131072, maxTokens: 131072,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"minimax-m2.1-free": { "minimax-m2.5-free": {
id: "minimax-m2.1-free", id: "minimax-m2.5-free",
name: "MiniMax M2.1 Free", name: "MiniMax M2.5 Free",
api: "anthropic-messages", api: "openai-completions",
provider: "opencode", provider: "opencode",
baseUrl: "https://opencode.ai/zen", baseUrl: "https://opencode.ai/zen/v1",
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
@ -5585,40 +5722,6 @@ export const MODELS = {
}, },
contextWindow: 204800, contextWindow: 204800,
maxTokens: 131072, maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
"qwen3-coder": {
id: "qwen3-coder",
name: "Qwen3 Coder",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.45,
output: 1.8,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"trinity-large-preview-free": {
id: "trinity-large-preview-free",
name: "Trinity Large Preview",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 131072,
maxTokens: 131072,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
}, },
"openrouter": { "openrouter": {
@ -5639,23 +5742,6 @@ export const MODELS = {
contextWindow: 256000, contextWindow: 256000,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"ai21/jamba-mini-1.7": {
id: "ai21/jamba-mini-1.7",
name: "AI21: Jamba Mini 1.7",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.19999999999999998,
output: 0.39999999999999997,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 256000,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"alibaba/tongyi-deepresearch-30b-a3b": { "alibaba/tongyi-deepresearch-30b-a3b": {
id: "alibaba/tongyi-deepresearch-30b-a3b", id: "alibaba/tongyi-deepresearch-30b-a3b",
name: "Tongyi DeepResearch 30B A3B", name: "Tongyi DeepResearch 30B A3B",
@ -6778,6 +6864,23 @@ export const MODELS = {
contextWindow: 196608, contextWindow: 196608,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"minimax/minimax-m2.5": {
id: "minimax/minimax-m2.5",
name: "MiniMax: MiniMax M2.5",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0.03,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"mistralai/codestral-2508": { "mistralai/codestral-2508": {
id: "mistralai/codestral-2508", id: "mistralai/codestral-2508",
name: "Mistral: Codestral 2508", name: "Mistral: Codestral 2508",
@ -6863,23 +6966,6 @@ export const MODELS = {
contextWindow: 262144, contextWindow: 262144,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/ministral-3b": {
id: "mistralai/ministral-3b",
name: "Mistral: Ministral 3B",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.04,
output: 0.04,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 131072,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/ministral-3b-2512": { "mistralai/ministral-3b-2512": {
id: "mistralai/ministral-3b-2512", id: "mistralai/ministral-3b-2512",
name: "Mistral: Ministral 3 3B 2512", name: "Mistral: Ministral 3 3B 2512",
@ -6897,23 +6983,6 @@ export const MODELS = {
contextWindow: 131072, contextWindow: 131072,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/ministral-8b": {
id: "mistralai/ministral-8b",
name: "Mistral: Ministral 8B",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.09999999999999999,
output: 0.09999999999999999,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 131072,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/ministral-8b-2512": { "mistralai/ministral-8b-2512": {
id: "mistralai/ministral-8b-2512", id: "mistralai/ministral-8b-2512",
name: "Mistral: Ministral 3 8B 2512", name: "Mistral: Ministral 3 8B 2512",
@ -7152,23 +7221,6 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/mistral-tiny": {
id: "mistralai/mistral-tiny",
name: "Mistral Tiny",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.25,
output: 0.25,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 32768,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/mixtral-8x22b-instruct": { "mistralai/mixtral-8x22b-instruct": {
id: "mistralai/mixtral-8x22b-instruct", id: "mistralai/mixtral-8x22b-instruct",
name: "Mistral: Mixtral 8x22B Instruct", name: "Mistral: Mixtral 8x22B Instruct",
@ -7203,23 +7255,6 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 16384, maxTokens: 16384,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/pixtral-12b": {
id: "mistralai/pixtral-12b",
name: "Mistral: Pixtral 12B",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text", "image"],
cost: {
input: 0.09999999999999999,
output: 0.09999999999999999,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 32768,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/pixtral-large-2411": { "mistralai/pixtral-large-2411": {
id: "mistralai/pixtral-large-2411", id: "mistralai/pixtral-large-2411",
name: "Mistral: Pixtral Large 2411", name: "Mistral: Pixtral Large 2411",
@ -7331,9 +7366,9 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text", "image"], input: ["text", "image"],
cost: { cost: {
input: 0.3, input: 0.44999999999999996,
output: 1.5, output: 2.25,
cacheRead: 0.049999999999999996, cacheRead: 0.070000002,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 262144, contextWindow: 262144,
@ -8359,6 +8394,23 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 100000, maxTokens: 100000,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"openrouter/aurora-alpha": {
id: "openrouter/aurora-alpha",
name: "Aurora Alpha",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 50000,
} satisfies Model<"openai-completions">,
"openrouter/auto": { "openrouter/auto": {
id: "openrouter/auto", id: "openrouter/auto",
name: "Auto Router", name: "Auto Router",
@ -8393,23 +8445,6 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"openrouter/pony-alpha": {
id: "openrouter/pony-alpha",
name: "Pony Alpha",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 200000,
maxTokens: 131000,
} satisfies Model<"openai-completions">,
"prime-intellect/intellect-3": { "prime-intellect/intellect-3": {
id: "prime-intellect/intellect-3", id: "prime-intellect/intellect-3",
name: "Prime Intellect: INTELLECT-3", name: "Prime Intellect: INTELLECT-3",
@ -8472,7 +8507,7 @@ export const MODELS = {
cost: { cost: {
input: 1.5999999999999999, input: 1.5999999999999999,
output: 6.3999999999999995, output: 6.3999999999999995,
cacheRead: 0.64, cacheRead: 0.32,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 32768, contextWindow: 32768,
@ -8489,11 +8524,11 @@ export const MODELS = {
cost: { cost: {
input: 0.39999999999999997, input: 0.39999999999999997,
output: 1.2, output: 1.2,
cacheRead: 0.16, cacheRead: 0.08,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 131072, contextWindow: 1000000,
maxTokens: 8192, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen-plus-2025-07-28": { "qwen/qwen-plus-2025-07-28": {
id: "qwen/qwen-plus-2025-07-28", id: "qwen/qwen-plus-2025-07-28",
@ -8522,7 +8557,7 @@ export const MODELS = {
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.39999999999999997, input: 0.39999999999999997,
output: 4, output: 1.2,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
@ -8540,10 +8575,10 @@ export const MODELS = {
cost: { cost: {
input: 0.049999999999999996, input: 0.049999999999999996,
output: 0.19999999999999998, output: 0.19999999999999998,
cacheRead: 0.02, cacheRead: 0.01,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 1000000, contextWindow: 131072,
maxTokens: 8192, maxTokens: 8192,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen-vl-max": { "qwen/qwen-vl-max": {
@ -8561,7 +8596,7 @@ export const MODELS = {
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 131072, contextWindow: 131072,
maxTokens: 8192, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-14b": { "qwen/qwen3-14b": {
id: "qwen/qwen3-14b", id: "qwen/qwen3-14b",
@ -8589,13 +8624,13 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.19999999999999998, input: 0.3,
output: 0.6, output: 1.2,
cacheRead: 0, cacheRead: 0.15,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 40960, contextWindow: 40960,
maxTokens: 4096, maxTokens: 40960,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-235b-a22b-2507": { "qwen/qwen3-235b-a22b-2507": {
id: "qwen/qwen3-235b-a22b-2507", id: "qwen/qwen3-235b-a22b-2507",
@ -8739,7 +8774,7 @@ export const MODELS = {
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
reasoning: true, reasoning: false,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.22, input: 0.22,
@ -8778,10 +8813,10 @@ export const MODELS = {
cost: { cost: {
input: 0.3, input: 0.3,
output: 1.5, output: 1.5,
cacheRead: 0.08, cacheRead: 0.06,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 1000000,
maxTokens: 65536, maxTokens: 65536,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-coder-next": { "qwen/qwen3-coder-next": {
@ -8812,10 +8847,10 @@ export const MODELS = {
cost: { cost: {
input: 1, input: 1,
output: 5, output: 5,
cacheRead: 0.09999999999999999, cacheRead: 0.19999999999999998,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 128000, contextWindow: 1000000,
maxTokens: 65536, maxTokens: 65536,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-coder:exacto": { "qwen/qwen3-coder:exacto": {
@ -8866,8 +8901,25 @@ export const MODELS = {
cacheRead: 0.24, cacheRead: 0.24,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 256000, contextWindow: 262144,
maxTokens: 32768, maxTokens: 65536,
} satisfies Model<"openai-completions">,
"qwen/qwen3-max-thinking": {
id: "qwen/qwen3-max-thinking",
name: "Qwen: Qwen3 Max Thinking",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text"],
cost: {
input: 1.2,
output: 6,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 65536,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-next-80b-a3b-instruct": { "qwen/qwen3-next-80b-a3b-instruct": {
id: "qwen/qwen3-next-80b-a3b-instruct", id: "qwen/qwen3-next-80b-a3b-instruct",
@ -8946,13 +8998,13 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text", "image"], input: ["text", "image"],
cost: { cost: {
input: 0.44999999999999996, input: 0,
output: 3.5, output: 0,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 262144, contextWindow: 131072,
maxTokens: 262144, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-vl-30b-a3b-instruct": { "qwen/qwen3-vl-30b-a3b-instruct": {
id: "qwen/qwen3-vl-30b-a3b-instruct", id: "qwen/qwen3-vl-30b-a3b-instruct",
@ -8963,13 +9015,13 @@ export const MODELS = {
reasoning: false, reasoning: false,
input: ["text", "image"], input: ["text", "image"],
cost: { cost: {
input: 0.15, input: 0.13,
output: 0.6, output: 0.52,
cacheRead: 0.075, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 262144, contextWindow: 131072,
maxTokens: 4096, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-vl-30b-a3b-thinking": { "qwen/qwen3-vl-30b-a3b-thinking": {
id: "qwen/qwen3-vl-30b-a3b-thinking", id: "qwen/qwen3-vl-30b-a3b-thinking",
@ -8980,8 +9032,25 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text", "image"], input: ["text", "image"],
cost: { cost: {
input: 0.19999999999999998, input: 0,
output: 1, output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 131072,
maxTokens: 32768,
} satisfies Model<"openai-completions">,
"qwen/qwen3-vl-32b-instruct": {
id: "qwen/qwen3-vl-32b-instruct",
name: "Qwen: Qwen3 VL 32B Instruct",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text", "image"],
cost: {
input: 0.10400000000000001,
output: 0.41600000000000004,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
@ -9014,12 +9083,12 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text", "image"], input: ["text", "image"],
cost: { cost: {
input: 0.18, input: 0.117,
output: 2.0999999999999996, output: 1.365,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 256000, contextWindow: 131072,
maxTokens: 32768, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwq-32b": { "qwen/qwq-32b": {
@ -9090,22 +9159,22 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 32768, maxTokens: 32768,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"stepfun-ai/step3": { "stepfun/step-3.5-flash": {
id: "stepfun-ai/step3", id: "stepfun/step-3.5-flash",
name: "StepFun: Step3", name: "StepFun: Step 3.5 Flash",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
reasoning: true, reasoning: true,
input: ["text", "image"], input: ["text"],
cost: { cost: {
input: 0.5700000000000001, input: 0.09999999999999999,
output: 1.42, output: 0.3,
cacheRead: 0, cacheRead: 0.02,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 65536, contextWindow: 256000,
maxTokens: 65536, maxTokens: 256000,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"stepfun/step-3.5-flash:free": { "stepfun/step-3.5-flash:free": {
id: "stepfun/step-3.5-flash:free", id: "stepfun/step-3.5-flash:free",
@ -9381,7 +9450,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4-32b": { "z-ai/glm-4-32b": {
id: "z-ai/glm-4-32b", id: "z-ai/glm-4-32b",
name: "Z.AI: GLM 4 32B ", name: "Z.ai: GLM 4 32B ",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9398,7 +9467,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.5": { "z-ai/glm-4.5": {
id: "z-ai/glm-4.5", id: "z-ai/glm-4.5",
name: "Z.AI: GLM 4.5", name: "Z.ai: GLM 4.5",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9415,7 +9484,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.5-air": { "z-ai/glm-4.5-air": {
id: "z-ai/glm-4.5-air", id: "z-ai/glm-4.5-air",
name: "Z.AI: GLM 4.5 Air", name: "Z.ai: GLM 4.5 Air",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9432,7 +9501,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.5-air:free": { "z-ai/glm-4.5-air:free": {
id: "z-ai/glm-4.5-air:free", id: "z-ai/glm-4.5-air:free",
name: "Z.AI: GLM 4.5 Air (free)", name: "Z.ai: GLM 4.5 Air (free)",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9449,7 +9518,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.5v": { "z-ai/glm-4.5v": {
id: "z-ai/glm-4.5v", id: "z-ai/glm-4.5v",
name: "Z.AI: GLM 4.5V", name: "Z.ai: GLM 4.5V",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9466,7 +9535,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.6": { "z-ai/glm-4.6": {
id: "z-ai/glm-4.6", id: "z-ai/glm-4.6",
name: "Z.AI: GLM 4.6", name: "Z.ai: GLM 4.6",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9483,7 +9552,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.6:exacto": { "z-ai/glm-4.6:exacto": {
id: "z-ai/glm-4.6:exacto", id: "z-ai/glm-4.6:exacto",
name: "Z.AI: GLM 4.6 (exacto)", name: "Z.ai: GLM 4.6 (exacto)",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9500,7 +9569,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.6v": { "z-ai/glm-4.6v": {
id: "z-ai/glm-4.6v", id: "z-ai/glm-4.6v",
name: "Z.AI: GLM 4.6V", name: "Z.ai: GLM 4.6V",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9517,7 +9586,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.7": { "z-ai/glm-4.7": {
id: "z-ai/glm-4.7", id: "z-ai/glm-4.7",
name: "Z.AI: GLM 4.7", name: "Z.ai: GLM 4.7",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9534,7 +9603,7 @@ export const MODELS = {
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-4.7-flash": { "z-ai/glm-4.7-flash": {
id: "z-ai/glm-4.7-flash", id: "z-ai/glm-4.7-flash",
name: "Z.AI: GLM 4.7 Flash", name: "Z.ai: GLM 4.7 Flash",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
@ -9549,6 +9618,23 @@ export const MODELS = {
contextWindow: 202752, contextWindow: 202752,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"z-ai/glm-5": {
id: "z-ai/glm-5",
name: "Z.ai: GLM 5",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0.7999999999999999,
output: 2.56,
cacheRead: 0.16,
cacheWrite: 0,
},
contextWindow: 202752,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
}, },
"vercel-ai-gateway": { "vercel-ai-gateway": {
"alibaba/qwen-3-14b": { "alibaba/qwen-3-14b": {
@ -10036,8 +10122,8 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.3, input: 0.21,
output: 1, output: 0.7899999999999999,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
@ -10061,22 +10147,22 @@ export const MODELS = {
contextWindow: 131072, contextWindow: 131072,
maxTokens: 65536, maxTokens: 65536,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"deepseek/deepseek-v3.2-exp": { "deepseek/deepseek-v3.2": {
id: "deepseek/deepseek-v3.2-exp", id: "deepseek/deepseek-v3.2",
name: "DeepSeek V3.2 Exp", name: "DeepSeek V3.2",
api: "anthropic-messages", api: "anthropic-messages",
provider: "vercel-ai-gateway", provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh", baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true, reasoning: false,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.27, input: 0.26,
output: 0.39999999999999997, output: 0.38,
cacheRead: 0, cacheRead: 0.13,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 163840, contextWindow: 128000,
maxTokens: 163840, maxTokens: 8000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"deepseek/deepseek-v3.2-thinking": { "deepseek/deepseek-v3.2-thinking": {
id: "deepseek/deepseek-v3.2-thinking", id: "deepseek/deepseek-v3.2-thinking",
@ -10435,6 +10521,23 @@ export const MODELS = {
contextWindow: 204800, contextWindow: 204800,
maxTokens: 131072, maxTokens: 131072,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"minimax/minimax-m2.5": {
id: "minimax/minimax-m2.5",
name: "MiniMax M2.5",
api: "anthropic-messages",
provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true,
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0.03,
cacheWrite: 0.375,
},
contextWindow: 204800,
maxTokens: 131000,
} satisfies Model<"anthropic-messages">,
"mistral/codestral": { "mistral/codestral": {
id: "mistral/codestral", id: "mistral/codestral",
name: "Mistral Codestral", name: "Mistral Codestral",
@ -10956,7 +11059,7 @@ export const MODELS = {
cost: { cost: {
input: 1.25, input: 1.25,
output: 10, output: 10,
cacheRead: 0.125, cacheRead: 0.13,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 400000, contextWindow: 400000,
@ -11676,6 +11779,23 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 128000, maxTokens: 128000,
} satisfies Model<"anthropic-messages">, } satisfies Model<"anthropic-messages">,
"zai/glm-5": {
id: "zai/glm-5",
name: "GLM-5",
api: "anthropic-messages",
provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true,
input: ["text"],
cost: {
input: 1,
output: 3.1999999999999997,
cacheRead: 0.19999999999999998,
cacheWrite: 0,
},
contextWindow: 202800,
maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
}, },
"xai": { "xai": {
"grok-2": { "grok-2": {
@ -12198,5 +12318,23 @@ export const MODELS = {
contextWindow: 200000, contextWindow: 200000,
maxTokens: 131072, maxTokens: 131072,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"glm-5": {
id: "glm-5",
name: "GLM-5",
api: "openai-completions",
provider: "zai",
baseUrl: "https://api.z.ai/api/coding/paas/v4",
compat: {"supportsDeveloperRole":false,"thinkingFormat":"zai"},
reasoning: true,
input: ["text"],
cost: {
input: 1,
output: 3.2,
cacheRead: 0.2,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
}, },
} as const; } as const;

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
### New Features ### New Features

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
## [0.52.9] - 2026-02-08 ## [0.52.9] - 2026-02-08

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
### Added ### Added

View file

@ -1,5 +1,7 @@
# Changelog # Changelog
## [Unreleased]
## [0.52.10] - 2026-02-12 ## [0.52.10] - 2026-02-12
### Fixed ### Fixed