Add [Unreleased] section for next cycle

This commit is contained in:
Mario Zechner 2026-02-17 20:04:38 +01:00
parent ce1410b01d
commit 4ba3e5be22
7 changed files with 301 additions and 31 deletions

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
## [0.52.12] - 2026-02-13

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
### Added

View file

@ -430,6 +430,23 @@ export const MODELS = {
contextWindow: 163840,
maxTokens: 81920,
} satisfies Model<"bedrock-converse-stream">,
"deepseek.v3.2-v1:0": {
id: "deepseek.v3.2-v1:0",
name: "DeepSeek-V3.2",
api: "bedrock-converse-stream",
provider: "amazon-bedrock",
baseUrl: "https://bedrock-runtime.us-east-1.amazonaws.com",
reasoning: true,
input: ["text"],
cost: {
input: 0.62,
output: 1.85,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 163840,
maxTokens: 81920,
} satisfies Model<"bedrock-converse-stream">,
"eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
id: "eu.anthropic.claude-haiku-4-5-20251001-v1:0",
name: "Claude Haiku 4.5 (EU)",
@ -1707,6 +1724,23 @@ export const MODELS = {
contextWindow: 200000,
maxTokens: 64000,
} satisfies Model<"anthropic-messages">,
"claude-sonnet-4-6": {
id: "claude-sonnet-4-6",
name: "Claude Sonnet 4.6",
api: "anthropic-messages",
provider: "anthropic",
baseUrl: "https://api.anthropic.com",
reasoning: true,
input: ["text", "image"],
cost: {
input: 3,
output: 15,
cacheRead: 0.3,
cacheWrite: 3.75,
},
contextWindow: 200000,
maxTokens: 64000,
} satisfies Model<"anthropic-messages">,
},
"azure-openai-responses": {
"codex-mini-latest": {
@ -2383,8 +2417,8 @@ export const MODELS = {
reasoning: false,
input: ["text"],
cost: {
input: 0,
output: 0,
input: 2.25,
output: 2.75,
cacheRead: 0,
cacheWrite: 0,
},
@ -3774,6 +3808,24 @@ export const MODELS = {
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"MiniMaxAI/MiniMax-M2.5": {
id: "MiniMaxAI/MiniMax-M2.5",
name: "MiniMax-M2.5",
api: "openai-completions",
provider: "huggingface",
baseUrl: "https://router.huggingface.co/v1",
compat: {"supportsDeveloperRole":false},
reasoning: true,
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0.03,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"Qwen/Qwen3-235B-A22B-Thinking-2507": {
id: "Qwen/Qwen3-235B-A22B-Thinking-2507",
name: "Qwen3-235B-A22B-Thinking-2507",
@ -3810,6 +3862,24 @@ export const MODELS = {
contextWindow: 262144,
maxTokens: 66536,
} satisfies Model<"openai-completions">,
"Qwen/Qwen3-Coder-Next": {
id: "Qwen/Qwen3-Coder-Next",
name: "Qwen3-Coder-Next",
api: "openai-completions",
provider: "huggingface",
baseUrl: "https://router.huggingface.co/v1",
compat: {"supportsDeveloperRole":false},
reasoning: false,
input: ["text"],
cost: {
input: 0.2,
output: 1.5,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"Qwen/Qwen3-Next-80B-A3B-Instruct": {
id: "Qwen/Qwen3-Next-80B-A3B-Instruct",
name: "Qwen3-Next-80B-A3B-Instruct",
@ -3846,6 +3916,24 @@ export const MODELS = {
contextWindow: 262144,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"Qwen/Qwen3.5-397B-A17B": {
id: "Qwen/Qwen3.5-397B-A17B",
name: "Qwen3.5-397B-A17B",
api: "openai-completions",
provider: "huggingface",
baseUrl: "https://router.huggingface.co/v1",
compat: {"supportsDeveloperRole":false},
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.6,
output: 3.6,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 32768,
} satisfies Model<"openai-completions">,
"XiaomiMiMo/MiMo-V2-Flash": {
id: "XiaomiMiMo/MiMo-V2-Flash",
name: "MiMo-V2-Flash",
@ -4115,6 +4203,23 @@ export const MODELS = {
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
"MiniMax-M2.5-highspeed": {
id: "MiniMax-M2.5-highspeed",
name: "MiniMax-M2.5-highspeed",
api: "anthropic-messages",
provider: "minimax",
baseUrl: "https://api.minimax.io/anthropic",
reasoning: true,
input: ["text"],
cost: {
input: 0.6,
output: 2.4,
cacheRead: 0.06,
cacheWrite: 0.375,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
},
"minimax-cn": {
"MiniMax-M2": {
@ -4168,6 +4273,23 @@ export const MODELS = {
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
"MiniMax-M2.5-highspeed": {
id: "MiniMax-M2.5-highspeed",
name: "MiniMax-M2.5-highspeed",
api: "anthropic-messages",
provider: "minimax-cn",
baseUrl: "https://api.minimaxi.com/anthropic",
reasoning: true,
input: ["text"],
cost: {
input: 0.6,
output: 2.4,
cacheRead: 0.06,
cacheWrite: 0.375,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"anthropic-messages">,
},
"mistral": {
"codestral-latest": {
@ -5536,6 +5658,40 @@ export const MODELS = {
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"glm-5": {
id: "glm-5",
name: "GLM-5",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: true,
input: ["text"],
cost: {
input: 1,
output: 3.2,
cacheRead: 0.2,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"glm-5-free": {
id: "glm-5-free",
name: "GLM-5 Free",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"gpt-5": {
id: "gpt-5",
name: "GPT-5",
@ -5774,6 +5930,23 @@ export const MODELS = {
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"minimax-m2.5": {
id: "minimax-m2.5",
name: "MiniMax M2.5",
api: "openai-completions",
provider: "opencode",
baseUrl: "https://opencode.ai/zen/v1",
reasoning: true,
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0.06,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
"minimax-m2.5-free": {
id: "minimax-m2.5-free",
name: "MiniMax M2.5 Free",
@ -6133,6 +6306,23 @@ export const MODELS = {
contextWindow: 1000000,
maxTokens: 64000,
} satisfies Model<"openai-completions">,
"anthropic/claude-sonnet-4.6": {
id: "anthropic/claude-sonnet-4.6",
name: "Anthropic: Claude Sonnet 4.6",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text", "image"],
cost: {
input: 3,
output: 15,
cacheRead: 0.3,
cacheWrite: 3.75,
},
contextWindow: 1000000,
maxTokens: 128000,
} satisfies Model<"openai-completions">,
"arcee-ai/trinity-large-preview:free": {
id: "arcee-ai/trinity-large-preview:free",
name: "Arcee AI: Trinity Large Preview (free)",
@ -6448,13 +6638,13 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.25,
input: 0.26,
output: 0.38,
cacheRead: 0.125,
cacheRead: 0.13,
cacheWrite: 0,
},
contextWindow: 163840,
maxTokens: 65536,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"deepseek/deepseek-v3.2-exp": {
id: "deepseek/deepseek-v3.2-exp",
@ -6942,12 +7132,12 @@ export const MODELS = {
input: ["text"],
cost: {
input: 0.3,
output: 1.2,
cacheRead: 0.03,
output: 1.1,
cacheRead: 0.15,
cacheWrite: 0,
},
contextWindow: 204800,
maxTokens: 131072,
contextWindow: 196608,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"mistralai/codestral-2508": {
id: "mistralai/codestral-2508",
@ -7383,13 +7573,13 @@ export const MODELS = {
reasoning: false,
input: ["text"],
cost: {
input: 0.39,
output: 1.9,
cacheRead: 0.195,
input: 0.39999999999999997,
output: 2,
cacheRead: 0.15,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 262144,
contextWindow: 131072,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"moonshotai/kimi-k2-0905:exacto": {
id: "moonshotai/kimi-k2-0905:exacto",
@ -7434,13 +7624,13 @@ export const MODELS = {
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.44999999999999996,
output: 2.25,
cacheRead: 0.070000002,
input: 0.22999999999999998,
output: 3,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 4096,
maxTokens: 262144,
} satisfies Model<"openai-completions">,
"nex-agi/deepseek-v3.1-nex-n1": {
id: "nex-agi/deepseek-v3.1-nex-n1",
@ -8913,9 +9103,9 @@ export const MODELS = {
reasoning: false,
input: ["text"],
cost: {
input: 0.07,
output: 0.3,
cacheRead: 0.035,
input: 0.12,
output: 0.75,
cacheRead: 0.06,
cacheWrite: 0,
},
contextWindow: 262144,
@ -9176,6 +9366,40 @@ export const MODELS = {
contextWindow: 131072,
maxTokens: 32768,
} satisfies Model<"openai-completions">,
"qwen/qwen3.5-397b-a17b": {
id: "qwen/qwen3.5-397b-a17b",
name: "Qwen: Qwen3.5 397B A17B",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.15,
output: 1,
cacheRead: 0.15,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"qwen/qwen3.5-plus-02-15": {
id: "qwen/qwen3.5-plus-02-15",
name: "Qwen: Qwen3.5 Plus 2026-02-15",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.39999999999999997,
output: 2.4,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 1000000,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"qwen/qwq-32b": {
id: "qwen/qwq-32b",
name: "Qwen: QwQ 32B",
@ -9610,9 +9834,9 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.35,
output: 1.5,
cacheRead: 0.175,
input: 0.33999999999999997,
output: 1.7,
cacheRead: 0.16999999999999998,
cacheWrite: 0,
},
contextWindow: 202752,
@ -9661,9 +9885,9 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.39999999999999997,
output: 1.5,
cacheRead: 0.19999999999999998,
input: 0.38,
output: 1.7,
cacheRead: 0.19,
cacheWrite: 0,
},
contextWindow: 202752,
@ -9695,12 +9919,12 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.7999999999999999,
output: 2.56,
cacheRead: 0.16,
input: 0.3,
output: 2.5500000000000003,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 202752,
contextWindow: 204800,
maxTokens: 131072,
} satisfies Model<"openai-completions">,
},
@ -9909,6 +10133,23 @@ export const MODELS = {
contextWindow: 256000,
maxTokens: 256000,
} satisfies Model<"anthropic-messages">,
"alibaba/qwen3.5-plus": {
id: "alibaba/qwen3.5-plus",
name: "Qwen 3.5 Plus",
api: "anthropic-messages",
provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true,
input: ["text", "image"],
cost: {
input: 0.39999999999999997,
output: 2.4,
cacheRead: 0.04,
cacheWrite: 0.5,
},
contextWindow: 1000000,
maxTokens: 64000,
} satisfies Model<"anthropic-messages">,
"anthropic/claude-3-haiku": {
id: "anthropic/claude-3-haiku",
name: "Claude 3 Haiku",
@ -10113,6 +10354,23 @@ export const MODELS = {
contextWindow: 1000000,
maxTokens: 64000,
} satisfies Model<"anthropic-messages">,
"anthropic/claude-sonnet-4.6": {
id: "anthropic/claude-sonnet-4.6",
name: "Claude Sonnet 4.6",
api: "anthropic-messages",
provider: "vercel-ai-gateway",
baseUrl: "https://ai-gateway.vercel.sh",
reasoning: true,
input: ["text", "image"],
cost: {
input: 3,
output: 15,
cacheRead: 0.3,
cacheWrite: 3.75,
},
contextWindow: 1000000,
maxTokens: 64000,
} satisfies Model<"anthropic-messages">,
"arcee-ai/trinity-large-preview": {
id: "arcee-ai/trinity-large-preview",
name: "Trinity Large Preview",

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
### Breaking Changes

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
## [0.52.12] - 2026-02-13

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
## [0.52.12] - 2026-02-13

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.53.0] - 2026-02-17
## [0.52.12] - 2026-02-13