diff --git a/packages/ai/scripts/generate-models.ts b/packages/ai/scripts/generate-models.ts index 3e1130b0..5cfa335d 100644 --- a/packages/ai/scripts/generate-models.ts +++ b/packages/ai/scripts/generate-models.ts @@ -689,7 +689,8 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: true, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + // the Model type doesn't seem to support having extended-context costs, so I'm just using the pricing for <200k input + cost: { input: 2, output: 12, cacheRead: 0.2, cacheWrite: 2.375 }, contextWindow: 1048576, maxTokens: 65535, }, @@ -701,7 +702,8 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: true, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + // the Model type doesn't seem to support having extended-context costs, so I'm just using the pricing for <200k input + cost: { input: 2, output: 12, cacheRead: 0.2, cacheWrite: 2.375 }, contextWindow: 1048576, maxTokens: 65535, }, @@ -713,7 +715,7 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: true, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + cost: { input: 0.5, output: 3, cacheRead: 0.5, cacheWrite: 0 }, contextWindow: 1048576, maxTokens: 65535, }, @@ -725,7 +727,7 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: false, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + cost: { input: 3, output: 15, cacheRead: 0.3, cacheWrite: 3.75 }, contextWindow: 200000, maxTokens: 64000, }, @@ -737,7 +739,7 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: true, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + cost: { input: 3, output: 15, cacheRead: 0.3, cacheWrite: 3.75 }, contextWindow: 200000, maxTokens: 64000, }, @@ -749,7 +751,7 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: true, input: ["text", "image"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + cost: { input: 5, output: 25, cacheRead: 0.5, cacheWrite: 6.25 }, contextWindow: 200000, maxTokens: 64000, }, @@ -761,7 +763,7 @@ async function generateModels() { baseUrl: ANTIGRAVITY_ENDPOINT, reasoning: false, input: ["text"], - cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + cost: { input: 0.09, output: 0.36, cacheRead: 0, cacheWrite: 0 }, contextWindow: 131072, maxTokens: 32768, }, diff --git a/packages/ai/src/models.generated.ts b/packages/ai/src/models.generated.ts index 4ead53e2..cbe25fea 100644 --- a/packages/ai/src/models.generated.ts +++ b/packages/ai/src/models.generated.ts @@ -1156,10 +1156,10 @@ export const MODELS = { reasoning: true, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, + input: 5, + output: 25, + cacheRead: 0.5, + cacheWrite: 6.25, }, contextWindow: 200000, maxTokens: 64000, @@ -1173,10 +1173,10 @@ export const MODELS = { reasoning: false, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, + input: 3, + output: 15, + cacheRead: 0.3, + cacheWrite: 3.75, }, contextWindow: 200000, maxTokens: 64000, @@ -1190,10 +1190,10 @@ export const MODELS = { reasoning: true, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, + input: 3, + output: 15, + cacheRead: 0.3, + cacheWrite: 3.75, }, contextWindow: 200000, maxTokens: 64000, @@ -1207,9 +1207,9 @@ export const MODELS = { reasoning: true, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, + input: 0.5, + output: 3, + cacheRead: 0.5, cacheWrite: 0, }, contextWindow: 1048576, @@ -1224,10 +1224,10 @@ export const MODELS = { reasoning: true, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, + input: 2, + output: 12, + cacheRead: 0.2, + cacheWrite: 2.375, }, contextWindow: 1048576, maxTokens: 65535, @@ -1241,10 +1241,10 @@ export const MODELS = { reasoning: true, input: ["text", "image"], cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, + input: 2, + output: 12, + cacheRead: 0.2, + cacheWrite: 2.375, }, contextWindow: 1048576, maxTokens: 65535, @@ -1258,8 +1258,8 @@ export const MODELS = { reasoning: false, input: ["text"], cost: { - input: 0, - output: 0, + input: 0.09, + output: 0.36, cacheRead: 0, cacheWrite: 0, },