From 946efe4b4561b7ae8c0818d9f83d2434c0c24d17 Mon Sep 17 00:00:00 2001 From: Mario Zechner Date: Thu, 8 Jan 2026 03:16:59 +0100 Subject: [PATCH] fix(ai): detect context_length_exceeded overflow --- packages/ai/CHANGELOG.md | 2 ++ packages/ai/src/utils/overflow.ts | 2 +- 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/packages/ai/CHANGELOG.md b/packages/ai/CHANGELOG.md index f3950637..307a69b1 100644 --- a/packages/ai/CHANGELOG.md +++ b/packages/ai/CHANGELOG.md @@ -13,6 +13,8 @@ ### Fixed - Fixed OpenAI Codex context window from 400,000 to 272,000 tokens to match Codex CLI defaults and prevent 400 errors. ([#536](https://github.com/badlogic/pi-mono/pull/536) by [@ghoulr](https://github.com/ghoulr)) +- Fixed Codex SSE error events to surface message, code, and status. ([#551](https://github.com/badlogic/pi-mono/pull/551) by [@tmustier](https://github.com/tmustier)) +- Fixed context overflow detection for `context_length_exceeded` error codes. ## [0.37.8] - 2026-01-07 diff --git a/packages/ai/src/utils/overflow.ts b/packages/ai/src/utils/overflow.ts index 8a84e4a3..15c3c6cc 100644 --- a/packages/ai/src/utils/overflow.ts +++ b/packages/ai/src/utils/overflow.ts @@ -32,7 +32,7 @@ const OVERFLOW_PATTERNS = [ /exceeds the limit of \d+/i, // GitHub Copilot /exceeds the available context size/i, // llama.cpp server /greater than the context length/i, // LM Studio - /context length exceeded/i, // Generic fallback + /context[_ ]length[_ ]exceeded/i, // Generic fallback /too many tokens/i, // Generic fallback /token limit exceeded/i, // Generic fallback ];