From c4102c7b81be9c555a3a2c387bb34a41bcaf32f0 Mon Sep 17 00:00:00 2001 From: Mario Zechner Date: Wed, 12 Nov 2025 23:22:40 +0100 Subject: [PATCH] docs: clarify auto-compaction options and fix flicker claim - Expanded auto-compaction: switch to bigger model and either continue or summarize - Fixed planned feature: MORE flicker than Claude Code (not less) --- packages/coding-agent/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/packages/coding-agent/README.md b/packages/coding-agent/README.md index 326f42e7..d67f736e 100644 --- a/packages/coding-agent/README.md +++ b/packages/coding-agent/README.md @@ -460,10 +460,10 @@ Things that might happen eventually: - **Custom/local models**: Support for Ollama, llama.cpp, vLLM, SGLang, LM Studio via JSON config file - **Auto-compaction**: Currently, watch the context percentage at the bottom. When it approaches 80%, either: - Ask the agent to write a summary .md file you can load in a new session - - Switch to a model with bigger context (e.g., Gemini) using `/model` and have it summarize the session + - Switch to a model with bigger context (e.g., Gemini) using `/model` and either continue with that model, or let it summarize the session to a .md file to be loaded in a new session - **Message queuing**: Core engine supports it, just needs UI wiring - **Better RPC mode docs**: It works, you'll figure it out (see `test/rpc-example.ts`) -- **Less flicker than Claude Code**: One day... +- **More flicker than Claude Code**: One day... ## License