mirror of
https://github.com/getcompanion-ai/co-mono.git
synced 2026-04-20 13:02:15 +00:00
Add [Unreleased] section for next cycle
This commit is contained in:
parent
0aaa4ef5e4
commit
e46cc55fe0
7 changed files with 80 additions and 34 deletions
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
## [0.45.7] - 2026-01-13
|
## [0.45.7] - 2026-01-13
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|
|
||||||
|
|
@ -5216,23 +5216,6 @@ export const MODELS = {
|
||||||
contextWindow: 1048576,
|
contextWindow: 1048576,
|
||||||
maxTokens: 65536,
|
maxTokens: 65536,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
"google/gemini-2.5-flash-preview-09-2025": {
|
|
||||||
id: "google/gemini-2.5-flash-preview-09-2025",
|
|
||||||
name: "Google: Gemini 2.5 Flash Preview 09-2025",
|
|
||||||
api: "openai-completions",
|
|
||||||
provider: "openrouter",
|
|
||||||
baseUrl: "https://openrouter.ai/api/v1",
|
|
||||||
reasoning: true,
|
|
||||||
input: ["text", "image"],
|
|
||||||
cost: {
|
|
||||||
input: 0.3,
|
|
||||||
output: 2.5,
|
|
||||||
cacheRead: 0.075,
|
|
||||||
cacheWrite: 0.3833,
|
|
||||||
},
|
|
||||||
contextWindow: 1048576,
|
|
||||||
maxTokens: 65536,
|
|
||||||
} satisfies Model<"openai-completions">,
|
|
||||||
"google/gemini-2.5-pro": {
|
"google/gemini-2.5-pro": {
|
||||||
id: "google/gemini-2.5-pro",
|
id: "google/gemini-2.5-pro",
|
||||||
name: "Google: Gemini 2.5 Pro",
|
name: "Google: Gemini 2.5 Pro",
|
||||||
|
|
@ -5981,6 +5964,23 @@ export const MODELS = {
|
||||||
contextWindow: 131072,
|
contextWindow: 131072,
|
||||||
maxTokens: 131072,
|
maxTokens: 131072,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
|
"mistralai/mistral-small-3.1-24b-instruct:free": {
|
||||||
|
id: "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||||
|
name: "Mistral: Mistral Small 3.1 24B (free)",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: false,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: {
|
||||||
|
input: 0,
|
||||||
|
output: 0,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 128000,
|
||||||
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
"mistralai/mistral-small-3.2-24b-instruct": {
|
"mistralai/mistral-small-3.2-24b-instruct": {
|
||||||
id: "mistralai/mistral-small-3.2-24b-instruct",
|
id: "mistralai/mistral-small-3.2-24b-instruct",
|
||||||
name: "Mistral: Mistral Small 3.2 24B",
|
name: "Mistral: Mistral Small 3.2 24B",
|
||||||
|
|
@ -7528,6 +7528,23 @@ export const MODELS = {
|
||||||
contextWindow: 40960,
|
contextWindow: 40960,
|
||||||
maxTokens: 40960,
|
maxTokens: 40960,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
|
"qwen/qwen3-4b:free": {
|
||||||
|
id: "qwen/qwen3-4b:free",
|
||||||
|
name: "Qwen: Qwen3 4B (free)",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: true,
|
||||||
|
input: ["text"],
|
||||||
|
cost: {
|
||||||
|
input: 0,
|
||||||
|
output: 0,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 40960,
|
||||||
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
"qwen/qwen3-8b": {
|
"qwen/qwen3-8b": {
|
||||||
id: "qwen/qwen3-8b",
|
id: "qwen/qwen3-8b",
|
||||||
name: "Qwen: Qwen3 8B",
|
name: "Qwen: Qwen3 8B",
|
||||||
|
|
@ -7630,6 +7647,23 @@ export const MODELS = {
|
||||||
contextWindow: 262144,
|
contextWindow: 262144,
|
||||||
maxTokens: 65536,
|
maxTokens: 65536,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
|
"qwen/qwen3-coder:free": {
|
||||||
|
id: "qwen/qwen3-coder:free",
|
||||||
|
name: "Qwen: Qwen3 Coder 480B A35B (free)",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: false,
|
||||||
|
input: ["text"],
|
||||||
|
cost: {
|
||||||
|
input: 0,
|
||||||
|
output: 0,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 262000,
|
||||||
|
maxTokens: 262000,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
"qwen/qwen3-max": {
|
"qwen/qwen3-max": {
|
||||||
id: "qwen/qwen3-max",
|
id: "qwen/qwen3-max",
|
||||||
name: "Qwen: Qwen3 Max",
|
name: "Qwen: Qwen3 Max",
|
||||||
|
|
@ -8089,6 +8123,23 @@ export const MODELS = {
|
||||||
contextWindow: 256000,
|
contextWindow: 256000,
|
||||||
maxTokens: 10000,
|
maxTokens: 10000,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
|
"xiaomi/mimo-v2-flash": {
|
||||||
|
id: "xiaomi/mimo-v2-flash",
|
||||||
|
name: "Xiaomi: MiMo-V2-Flash",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: true,
|
||||||
|
input: ["text"],
|
||||||
|
cost: {
|
||||||
|
input: 0.09999999999999999,
|
||||||
|
output: 0.3,
|
||||||
|
cacheRead: 0.02,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 262144,
|
||||||
|
maxTokens: 32000,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
"xiaomi/mimo-v2-flash:free": {
|
"xiaomi/mimo-v2-flash:free": {
|
||||||
id: "xiaomi/mimo-v2-flash:free",
|
id: "xiaomi/mimo-v2-flash:free",
|
||||||
name: "Xiaomi: MiMo-V2-Flash (free)",
|
name: "Xiaomi: MiMo-V2-Flash (free)",
|
||||||
|
|
@ -9247,23 +9298,6 @@ export const MODELS = {
|
||||||
contextWindow: 128000,
|
contextWindow: 128000,
|
||||||
maxTokens: 64000,
|
maxTokens: 64000,
|
||||||
} satisfies Model<"anthropic-messages">,
|
} satisfies Model<"anthropic-messages">,
|
||||||
"mistral/mistral-nemo": {
|
|
||||||
id: "mistral/mistral-nemo",
|
|
||||||
name: "Mistral Nemo",
|
|
||||||
api: "anthropic-messages",
|
|
||||||
provider: "vercel-ai-gateway",
|
|
||||||
baseUrl: "https://ai-gateway.vercel.sh",
|
|
||||||
reasoning: false,
|
|
||||||
input: ["text"],
|
|
||||||
cost: {
|
|
||||||
input: 0.04,
|
|
||||||
output: 0.16999999999999998,
|
|
||||||
cacheRead: 0,
|
|
||||||
cacheWrite: 0,
|
|
||||||
},
|
|
||||||
contextWindow: 60288,
|
|
||||||
maxTokens: 16000,
|
|
||||||
} satisfies Model<"anthropic-messages">,
|
|
||||||
"mistral/mistral-small": {
|
"mistral/mistral-small": {
|
||||||
id: "mistral/mistral-small",
|
id: "mistral/mistral-small",
|
||||||
name: "Mistral Small",
|
name: "Mistral Small",
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
### Fixed
|
### Fixed
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
## [0.45.7] - 2026-01-13
|
## [0.45.7] - 2026-01-13
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
### Fixed
|
### Fixed
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.46.0] - 2026-01-15
|
## [0.46.0] - 2026-01-15
|
||||||
|
|
||||||
## [0.45.7] - 2026-01-13
|
## [0.45.7] - 2026-01-13
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue