mirror of
https://github.com/getcompanion-ai/co-mono.git
synced 2026-04-20 17:02:11 +00:00
Add [Unreleased] section
This commit is contained in:
parent
b193560ab0
commit
6a29b2df3f
2 changed files with 90 additions and 37 deletions
|
|
@ -1080,6 +1080,23 @@ export const MODELS = {
|
||||||
contextWindow: 128000,
|
contextWindow: 128000,
|
||||||
maxTokens: 16384,
|
maxTokens: 16384,
|
||||||
} satisfies Model<"openai-responses">,
|
} satisfies Model<"openai-responses">,
|
||||||
|
"gpt-5.1-codex-max": {
|
||||||
|
id: "gpt-5.1-codex-max",
|
||||||
|
name: "GPT-5.1 Codex Max",
|
||||||
|
api: "openai-responses",
|
||||||
|
provider: "openai",
|
||||||
|
baseUrl: "https://api.openai.com/v1",
|
||||||
|
reasoning: true,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: {
|
||||||
|
input: 1.25,
|
||||||
|
output: 10,
|
||||||
|
cacheRead: 0.125,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 400000,
|
||||||
|
maxTokens: 128000,
|
||||||
|
} satisfies Model<"openai-responses">,
|
||||||
o3: {
|
o3: {
|
||||||
id: "o3",
|
id: "o3",
|
||||||
name: "o3",
|
name: "o3",
|
||||||
|
|
@ -1199,23 +1216,6 @@ export const MODELS = {
|
||||||
contextWindow: 128000,
|
contextWindow: 128000,
|
||||||
maxTokens: 16384,
|
maxTokens: 16384,
|
||||||
} satisfies Model<"openai-responses">,
|
} satisfies Model<"openai-responses">,
|
||||||
"gpt-5.1-codex-max": {
|
|
||||||
id: "gpt-5.1-codex-max",
|
|
||||||
name: "GPT-5.1 Codex Max",
|
|
||||||
api: "openai-responses",
|
|
||||||
provider: "openai",
|
|
||||||
baseUrl: "https://api.openai.com/v1",
|
|
||||||
reasoning: true,
|
|
||||||
input: ["text", "image"],
|
|
||||||
cost: {
|
|
||||||
input: 1.25,
|
|
||||||
output: 10,
|
|
||||||
cacheRead: 0.125,
|
|
||||||
cacheWrite: 0,
|
|
||||||
},
|
|
||||||
contextWindow: 400000,
|
|
||||||
maxTokens: 128000,
|
|
||||||
} satisfies Model<"openai-responses">,
|
|
||||||
},
|
},
|
||||||
groq: {
|
groq: {
|
||||||
"llama-3.1-8b-instant": {
|
"llama-3.1-8b-instant": {
|
||||||
|
|
@ -2025,6 +2025,57 @@ export const MODELS = {
|
||||||
contextWindow: 1000000,
|
contextWindow: 1000000,
|
||||||
maxTokens: 65535,
|
maxTokens: 65535,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
|
"mistralai/ministral-14b-2512": {
|
||||||
|
id: "mistralai/ministral-14b-2512",
|
||||||
|
name: "Mistral: Ministral 3 14B 2512",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: false,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: {
|
||||||
|
input: 0.19999999999999998,
|
||||||
|
output: 0.19999999999999998,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 262144,
|
||||||
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
|
"mistralai/ministral-8b-2512": {
|
||||||
|
id: "mistralai/ministral-8b-2512",
|
||||||
|
name: "Mistral: Ministral 3 8B 2512",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: false,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: {
|
||||||
|
input: 0.15,
|
||||||
|
output: 0.15,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 262144,
|
||||||
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
|
"mistralai/ministral-3b-2512": {
|
||||||
|
id: "mistralai/ministral-3b-2512",
|
||||||
|
name: "Mistral: Ministral 3 3B 2512",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: false,
|
||||||
|
input: ["text", "image"],
|
||||||
|
cost: {
|
||||||
|
input: 0.09999999999999999,
|
||||||
|
output: 0.09999999999999999,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 131072,
|
||||||
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
"mistralai/mistral-large-2512": {
|
"mistralai/mistral-large-2512": {
|
||||||
id: "mistralai/mistral-large-2512",
|
id: "mistralai/mistral-large-2512",
|
||||||
name: "Mistral: Mistral Large 3 2512",
|
name: "Mistral: Mistral Large 3 2512",
|
||||||
|
|
@ -3351,23 +3402,6 @@ export const MODELS = {
|
||||||
contextWindow: 400000,
|
contextWindow: 400000,
|
||||||
maxTokens: 128000,
|
maxTokens: 128000,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
"openai/gpt-oss-120b:exacto": {
|
|
||||||
id: "openai/gpt-oss-120b:exacto",
|
|
||||||
name: "OpenAI: gpt-oss-120b (exacto)",
|
|
||||||
api: "openai-completions",
|
|
||||||
provider: "openrouter",
|
|
||||||
baseUrl: "https://openrouter.ai/api/v1",
|
|
||||||
reasoning: true,
|
|
||||||
input: ["text"],
|
|
||||||
cost: {
|
|
||||||
input: 0.04,
|
|
||||||
output: 0.19999999999999998,
|
|
||||||
cacheRead: 0,
|
|
||||||
cacheWrite: 0,
|
|
||||||
},
|
|
||||||
contextWindow: 131072,
|
|
||||||
maxTokens: 32768,
|
|
||||||
} satisfies Model<"openai-completions">,
|
|
||||||
"openai/gpt-oss-120b": {
|
"openai/gpt-oss-120b": {
|
||||||
id: "openai/gpt-oss-120b",
|
id: "openai/gpt-oss-120b",
|
||||||
name: "OpenAI: gpt-oss-120b",
|
name: "OpenAI: gpt-oss-120b",
|
||||||
|
|
@ -3377,13 +3411,30 @@ export const MODELS = {
|
||||||
reasoning: true,
|
reasoning: true,
|
||||||
input: ["text"],
|
input: ["text"],
|
||||||
cost: {
|
cost: {
|
||||||
input: 0.04,
|
input: 0.039,
|
||||||
output: 0.19999999999999998,
|
output: 0.19,
|
||||||
cacheRead: 0,
|
cacheRead: 0,
|
||||||
cacheWrite: 0,
|
cacheWrite: 0,
|
||||||
},
|
},
|
||||||
contextWindow: 131072,
|
contextWindow: 131072,
|
||||||
maxTokens: 32768,
|
maxTokens: 4096,
|
||||||
|
} satisfies Model<"openai-completions">,
|
||||||
|
"openai/gpt-oss-120b:exacto": {
|
||||||
|
id: "openai/gpt-oss-120b:exacto",
|
||||||
|
name: "OpenAI: gpt-oss-120b (exacto)",
|
||||||
|
api: "openai-completions",
|
||||||
|
provider: "openrouter",
|
||||||
|
baseUrl: "https://openrouter.ai/api/v1",
|
||||||
|
reasoning: true,
|
||||||
|
input: ["text"],
|
||||||
|
cost: {
|
||||||
|
input: 0.039,
|
||||||
|
output: 0.19,
|
||||||
|
cacheRead: 0,
|
||||||
|
cacheWrite: 0,
|
||||||
|
},
|
||||||
|
contextWindow: 131072,
|
||||||
|
maxTokens: 4096,
|
||||||
} satisfies Model<"openai-completions">,
|
} satisfies Model<"openai-completions">,
|
||||||
"openai/gpt-oss-20b:free": {
|
"openai/gpt-oss-20b:free": {
|
||||||
id: "openai/gpt-oss-20b:free",
|
id: "openai/gpt-oss-20b:free",
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,7 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [Unreleased]
|
||||||
|
|
||||||
## [0.12.11] - 2025-12-05
|
## [0.12.11] - 2025-12-05
|
||||||
|
|
||||||
### Changed
|
### Changed
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue