Merge pull request #281 from getcompanion-ai/codex5.3

model upgrade
This commit is contained in:
Hari 2026-03-09 19:38:00 -04:00 committed by GitHub
commit a018a0cd3e
5 changed files with 12 additions and 9 deletions

View file

@ -318,7 +318,7 @@ const authStorage = AuthStorage.create();
const modelRegistry = new ModelRegistry(authStorage); const modelRegistry = new ModelRegistry(authStorage);
// Find specific built-in model (doesn't check if API key exists) // Find specific built-in model (doesn't check if API key exists)
const opus = getModel("anthropic", "claude-opus-4-5"); const opus = getModel("anthropic", "claude-opus-4-6");
if (!opus) throw new Error("Model not found"); if (!opus) throw new Error("Model not found");
// Find any model by provider/id, including custom models from models.json // Find any model by provider/id, including custom models from models.json
@ -841,7 +841,7 @@ const statusTool: ToolDefinition = {
}), }),
}; };
const model = getModel("anthropic", "claude-opus-4-5"); const model = getModel("anthropic", "claude-opus-4-6");
if (!model) throw new Error("Model not found"); if (!model) throw new Error("Model not found");
// In-memory settings with overrides // In-memory settings with overrides

View file

@ -21,7 +21,7 @@ export const defaultModelPerProvider: Record<KnownProvider, string> = {
anthropic: "claude-opus-4-6", anthropic: "claude-opus-4-6",
openai: "gpt-5.4", openai: "gpt-5.4",
"azure-openai-responses": "gpt-5.2", "azure-openai-responses": "gpt-5.2",
"openai-codex": "gpt-5.4", "openai-codex": "gpt-5.3-codex",
google: "gemini-2.5-pro", google: "gemini-2.5-pro",
"google-gemini-cli": "gemini-2.5-pro", "google-gemini-cli": "gemini-2.5-pro",
"google-antigravity": "gemini-3.1-pro-high", "google-antigravity": "gemini-3.1-pro-high",

View file

@ -151,7 +151,7 @@ function getDefaultAgentDir(): string {
* // With explicit model * // With explicit model
* import { getModel } from '@mariozechner/pi-ai'; * import { getModel } from '@mariozechner/pi-ai';
* const { session } = await createAgentSession({ * const { session } = await createAgentSession({
* model: getModel('anthropic', 'claude-opus-4-5'), * model: getModel('anthropic', 'claude-opus-4-6'),
* thinkingLevel: 'high', * thinkingLevel: 'high',
* }); * });
* *

View file

@ -395,9 +395,12 @@ describe("resolveCliModel", () => {
}); });
describe("default model selection", () => { describe("default model selection", () => {
test("openai defaults are gpt-5.4", () => { test("openai default stays on gpt-5.4", () => {
expect(defaultModelPerProvider.openai).toBe("gpt-5.4"); expect(defaultModelPerProvider.openai).toBe("gpt-5.4");
expect(defaultModelPerProvider["openai-codex"]).toBe("gpt-5.4"); });
test("openai-codex default is gpt-5.3-codex", () => {
expect(defaultModelPerProvider["openai-codex"]).toBe("gpt-5.3-codex");
}); });
test("ai-gateway default is opus 4.6", () => { test("ai-gateway default is opus 4.6", () => {

View file

@ -40,7 +40,7 @@ describe("SettingsManager", () => {
// Simulate user editing settings.json externally to add enabledModels // Simulate user editing settings.json externally to add enabledModels
const currentSettings = JSON.parse(readFileSync(settingsPath, "utf-8")); const currentSettings = JSON.parse(readFileSync(settingsPath, "utf-8"));
currentSettings.enabledModels = ["claude-opus-4-5", "gpt-5.2-codex"]; currentSettings.enabledModels = ["claude-opus-4-6", "gpt-5.3-codex"];
writeFileSync(settingsPath, JSON.stringify(currentSettings, null, 2)); writeFileSync(settingsPath, JSON.stringify(currentSettings, null, 2));
// User changes thinking level via Shift+Tab // User changes thinking level via Shift+Tab
@ -50,8 +50,8 @@ describe("SettingsManager", () => {
// Verify enabledModels is preserved // Verify enabledModels is preserved
const savedSettings = JSON.parse(readFileSync(settingsPath, "utf-8")); const savedSettings = JSON.parse(readFileSync(settingsPath, "utf-8"));
expect(savedSettings.enabledModels).toEqual([ expect(savedSettings.enabledModels).toEqual([
"claude-opus-4-5", "claude-opus-4-6",
"gpt-5.2-codex", "gpt-5.3-codex",
]); ]);
expect(savedSettings.defaultThinkingLevel).toBe("high"); expect(savedSettings.defaultThinkingLevel).toBe("high");
expect(savedSettings.theme).toBe("dark"); expect(savedSettings.theme).toBe("dark");