Fix token counts in google-gemini-cli.ts for anthropic models.

This commit is contained in:
Mario Zechner 2025-12-20 18:58:26 +01:00
parent 24f89cf070
commit 2a0283ecfd

View file

@ -335,12 +335,15 @@ export const streamGoogleGeminiCli: StreamFunction<"google-gemini-cli"> = (
} }
if (responseData.usageMetadata) { if (responseData.usageMetadata) {
// promptTokenCount includes cachedContentTokenCount, so subtract to get fresh input
const promptTokens = responseData.usageMetadata.promptTokenCount || 0;
const cacheReadTokens = responseData.usageMetadata.cachedContentTokenCount || 0;
output.usage = { output.usage = {
input: responseData.usageMetadata.promptTokenCount || 0, input: promptTokens - cacheReadTokens,
output: output:
(responseData.usageMetadata.candidatesTokenCount || 0) + (responseData.usageMetadata.candidatesTokenCount || 0) +
(responseData.usageMetadata.thoughtsTokenCount || 0), (responseData.usageMetadata.thoughtsTokenCount || 0),
cacheRead: responseData.usageMetadata.cachedContentTokenCount || 0, cacheRead: cacheReadTokens,
cacheWrite: 0, cacheWrite: 0,
totalTokens: responseData.usageMetadata.totalTokenCount || 0, totalTokens: responseData.usageMetadata.totalTokenCount || 0,
cost: { cost: {