Add [Unreleased] section for next cycle

This commit is contained in:
Mario Zechner 2026-01-17 22:02:59 +01:00
parent 3e6ae39083
commit 632495338f
7 changed files with 27 additions and 15 deletions

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
## [0.48.0] - 2026-01-16

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
### Changed

View file

@ -4986,13 +4986,13 @@ export const MODELS = {
reasoning: true,
input: ["text"],
cost: {
input: 0.44999999999999996,
output: 2.1500000000000004,
input: 0.39999999999999997,
output: 1.75,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 131072,
maxTokens: 32768,
contextWindow: 163840,
maxTokens: 65536,
} satisfies Model<"openai-completions">,
"deepseek/deepseek-r1-distill-llama-70b": {
id: "deepseek/deepseek-r1-distill-llama-70b",
@ -5091,7 +5091,7 @@ export const MODELS = {
input: 0.09999999999999999,
output: 0.39999999999999997,
cacheRead: 0.024999999999999998,
cacheWrite: 0.18330000000000002,
cacheWrite: 0.0833,
},
contextWindow: 1048576,
maxTokens: 8192,
@ -5142,7 +5142,7 @@ export const MODELS = {
input: 0.3,
output: 2.5,
cacheRead: 0.03,
cacheWrite: 0.3833,
cacheWrite: 0.08333333333333334,
},
contextWindow: 1048576,
maxTokens: 65535,
@ -5159,7 +5159,7 @@ export const MODELS = {
input: 0.09999999999999999,
output: 0.39999999999999997,
cacheRead: 0.01,
cacheWrite: 0.18330000000000002,
cacheWrite: 0.0833,
},
contextWindow: 1048576,
maxTokens: 65535,
@ -5176,7 +5176,7 @@ export const MODELS = {
input: 0.09999999999999999,
output: 0.39999999999999997,
cacheRead: 0.01,
cacheWrite: 1,
cacheWrite: 0.0833,
},
contextWindow: 1048576,
maxTokens: 65536,
@ -5193,7 +5193,7 @@ export const MODELS = {
input: 0.3,
output: 2.5,
cacheRead: 0.075,
cacheWrite: 0.3833,
cacheWrite: 0.0833,
},
contextWindow: 1048576,
maxTokens: 65535,
@ -5210,7 +5210,7 @@ export const MODELS = {
input: 1.25,
output: 10,
cacheRead: 0.125,
cacheWrite: 1.625,
cacheWrite: 0.375,
},
contextWindow: 1048576,
maxTokens: 65536,
@ -5227,7 +5227,7 @@ export const MODELS = {
input: 1.25,
output: 10,
cacheRead: 0.31,
cacheWrite: 1.625,
cacheWrite: 0.375,
},
contextWindow: 1048576,
maxTokens: 65536,
@ -5244,7 +5244,7 @@ export const MODELS = {
input: 1.25,
output: 10,
cacheRead: 0.31,
cacheWrite: 1.625,
cacheWrite: 0.375,
},
contextWindow: 1048576,
maxTokens: 65535,
@ -5278,7 +5278,7 @@ export const MODELS = {
input: 2,
output: 12,
cacheRead: 0.19999999999999998,
cacheWrite: 2.375,
cacheWrite: 0.375,
},
contextWindow: 1048576,
maxTokens: 65536,
@ -7643,8 +7643,8 @@ export const MODELS = {
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 262144,
maxTokens: 262144,
contextWindow: 128000,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"qwen/qwen3-vl-235b-a22b-instruct": {
id: "qwen/qwen3-vl-235b-a22b-instruct",

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
### Added

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
## [0.48.0] - 2026-01-16

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
### Added

View file

@ -1,5 +1,7 @@
# Changelog
## [Unreleased]
## [0.49.0] - 2026-01-17
## [0.48.0] - 2026-01-16