diff --git a/biome.json b/biome.json index e00394fc..95ba4216 100644 --- a/biome.json +++ b/biome.json @@ -35,6 +35,7 @@ "*.md", "!**/node_modules/**/*", "!**/test-sessions.ts", + "!**/models.generated.ts", "!packages/web-ui/src/app.css", "!packages/mom/data/**/*", "!!**/node_modules" diff --git a/package-lock.json b/package-lock.json index 5d47258a..949907c1 100644 --- a/package-lock.json +++ b/package-lock.json @@ -6063,11 +6063,11 @@ }, "packages/agent": { "name": "@mariozechner/pi-agent-core", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0" + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1" }, "devDependencies": { "@types/node": "^24.3.0", @@ -6097,7 +6097,7 @@ }, "packages/ai": { "name": "@mariozechner/pi-ai", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { "@anthropic-ai/sdk": "0.71.2", @@ -6139,12 +6139,12 @@ }, "packages/coding-agent": { "name": "@mariozechner/pi-coding-agent", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { - "@mariozechner/pi-agent-core": "^0.22.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1", "chalk": "^5.5.0", "diff": "^8.0.2", "glob": "^11.0.3", @@ -6182,13 +6182,13 @@ }, "packages/mom": { "name": "@mariozechner/pi-mom", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { "@anthropic-ai/sandbox-runtime": "^0.0.16", - "@mariozechner/pi-agent-core": "^0.22.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-coding-agent": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-coding-agent": "^0.22.1", "@sinclair/typebox": "^0.34.0", "@slack/socket-mode": "^2.0.0", "@slack/web-api": "^7.0.0", @@ -6227,10 +6227,10 @@ }, "packages/pods": { "name": "@mariozechner/pi", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { - "@mariozechner/pi-agent-core": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", "chalk": "^5.5.0" }, "bin": { @@ -6243,7 +6243,7 @@ }, "packages/proxy": { "name": "@mariozechner/pi-proxy", - "version": "0.22.0", + "version": "0.22.1", "dependencies": { "@hono/node-server": "^1.14.0", "hono": "^4.6.16" @@ -6259,7 +6259,7 @@ }, "packages/tui": { "name": "@mariozechner/pi-tui", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { "@types/mime-types": "^2.1.4", @@ -6303,12 +6303,12 @@ }, "packages/web-ui": { "name": "@mariozechner/pi-web-ui", - "version": "0.22.0", + "version": "0.22.1", "license": "MIT", "dependencies": { "@lmstudio/sdk": "^1.5.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1", "docx-preview": "^0.3.7", "jszip": "^3.10.1", "lucide": "^0.544.0", @@ -6329,7 +6329,7 @@ }, "packages/web-ui/example": { "name": "pi-web-ui-example", - "version": "1.10.0", + "version": "1.10.1", "dependencies": { "@mariozechner/mini-lit": "^0.2.0", "@mariozechner/pi-ai": "file:../../ai", diff --git a/packages/agent/package.json b/packages/agent/package.json index 6bc84030..fe292b6b 100644 --- a/packages/agent/package.json +++ b/packages/agent/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-agent-core", - "version": "0.22.0", + "version": "0.22.1", "description": "General-purpose agent with transport abstraction, state management, and attachment support", "type": "module", "main": "./dist/index.js", @@ -18,8 +18,8 @@ "prepublishOnly": "npm run clean && npm run build" }, "dependencies": { - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0" + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1" }, "keywords": [ "ai", diff --git a/packages/ai/CHANGELOG.md b/packages/ai/CHANGELOG.md index 7b972d46..8603e08a 100644 --- a/packages/ai/CHANGELOG.md +++ b/packages/ai/CHANGELOG.md @@ -2,6 +2,14 @@ ## [Unreleased] +## [0.22.1] - 2025-12-15 + +_Dedicated to Peter's shoulder ([@steipete](https://twitter.com/steipete))_ + +### Added + +- **Interleaved thinking for Anthropic**: Enabled interleaved thinking in the Anthropic provider, allowing Claude models to output thinking blocks interspersed with text responses. + ## [0.22.0] - 2025-12-15 ### Added diff --git a/packages/ai/package.json b/packages/ai/package.json index f95cb518..151bad90 100644 --- a/packages/ai/package.json +++ b/packages/ai/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-ai", - "version": "0.22.0", + "version": "0.22.1", "description": "Unified LLM API with automatic model discovery and provider configuration", "type": "module", "main": "./dist/index.js", diff --git a/packages/ai/src/models.generated.ts b/packages/ai/src/models.generated.ts index bdd6490f..6077c44c 100644 --- a/packages/ai/src/models.generated.ts +++ b/packages/ai/src/models.generated.ts @@ -4,7 +4,7 @@ import type { Model } from "./types.js"; export const MODELS = { - anthropic: { + "anthropic": { "claude-opus-4-0": { id: "claude-opus-4-0", name: "Claude Opus 4 (latest)", @@ -363,7 +363,7 @@ export const MODELS = { maxTokens: 64000, } satisfies Model<"anthropic-messages">, }, - google: { + "google": { "gemini-2.5-flash-preview-05-20": { id: "gemini-2.5-flash-preview-05-20", name: "Gemini 2.5 Flash Preview 05-20", @@ -705,7 +705,7 @@ export const MODELS = { maxTokens: 8192, } satisfies Model<"google-generative-ai">, }, - openai: { + "openai": { "gpt-4.1-nano": { id: "gpt-4.1-nano", name: "GPT-4.1 nano", @@ -1046,7 +1046,7 @@ export const MODELS = { contextWindow: 200000, maxTokens: 100000, } satisfies Model<"openai-responses">, - o1: { + "o1": { id: "o1", name: "o1", api: "openai-responses", @@ -1131,7 +1131,7 @@ export const MODELS = { contextWindow: 400000, maxTokens: 128000, } satisfies Model<"openai-responses">, - o3: { + "o3": { id: "o3", name: "o3", api: "openai-responses", @@ -1268,7 +1268,7 @@ export const MODELS = { maxTokens: 16384, } satisfies Model<"openai-responses">, }, - groq: { + "groq": { "llama-3.1-8b-instant": { id: "llama-3.1-8b-instant", name: "Llama 3.1 8B Instant", @@ -1525,7 +1525,7 @@ export const MODELS = { maxTokens: 8192, } satisfies Model<"openai-completions">, }, - cerebras: { + "cerebras": { "qwen-3-235b-a22b-instruct-2507": { id: "qwen-3-235b-a22b-instruct-2507", name: "Qwen 3 235B Instruct", @@ -1578,7 +1578,7 @@ export const MODELS = { maxTokens: 32768, } satisfies Model<"openai-completions">, }, - xai: { + "xai": { "grok-4-fast-non-reasoning": { id: "grok-4-fast-non-reasoning", name: "Grok 4 Fast (Non-Reasoning)", @@ -1954,7 +1954,7 @@ export const MODELS = { maxTokens: 8192, } satisfies Model<"openai-completions">, }, - zai: { + "zai": { "glm-4.5-flash": { id: "glm-4.5-flash", name: "GLM-4.5-Flash", @@ -2058,7 +2058,7 @@ export const MODELS = { maxTokens: 32768, } satisfies Model<"anthropic-messages">, }, - mistral: { + "mistral": { "devstral-medium-2507": { id: "devstral-medium-2507", name: "Devstral Medium", @@ -2475,13 +2475,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text"], cost: { @@ -2499,12 +2494,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2522,13 +2512,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2546,13 +2531,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2570,13 +2550,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2594,12 +2569,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2617,12 +2587,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2640,12 +2605,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2663,13 +2623,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: false, input: ["text", "image"], cost: { @@ -2687,13 +2642,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: false, input: ["text", "image"], cost: { @@ -2711,12 +2661,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2734,13 +2679,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: false, input: ["text", "image"], cost: { @@ -2758,12 +2698,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2781,13 +2716,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2805,12 +2735,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2828,13 +2753,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2852,12 +2772,7 @@ export const MODELS = { api: "openai-responses", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, reasoning: true, input: ["text", "image"], cost: { @@ -2875,13 +2790,8 @@ export const MODELS = { api: "openai-completions", provider: "github-copilot", baseUrl: "https://api.individual.githubcopilot.com", - headers: { - "User-Agent": "GitHubCopilotChat/0.35.0", - "Editor-Version": "vscode/1.107.0", - "Editor-Plugin-Version": "copilot-chat/0.35.0", - "Copilot-Integration-Id": "vscode-chat", - }, - compat: { supportsStore: false, supportsDeveloperRole: false, supportsReasoningEffort: false }, + headers: {"User-Agent":"GitHubCopilotChat/0.35.0","Editor-Version":"vscode/1.107.0","Editor-Plugin-Version":"copilot-chat/0.35.0","Copilot-Integration-Id":"vscode-chat"}, + compat: {"supportsStore":false,"supportsDeveloperRole":false,"supportsReasoningEffort":false}, reasoning: true, input: ["text", "image"], cost: { @@ -2894,7 +2804,7 @@ export const MODELS = { maxTokens: 16000, } satisfies Model<"openai-completions">, }, - openrouter: { + "openrouter": { "nvidia/nemotron-3-nano-30b-a3b:free": { id: "nvidia/nemotron-3-nano-30b-a3b:free", name: "NVIDIA: Nemotron 3 Nano 30B A3B (free)", @@ -6023,9 +5933,9 @@ export const MODELS = { contextWindow: 32768, maxTokens: 4096, } satisfies Model<"openai-completions">, - "anthropic/claude-3.5-haiku-20241022": { - id: "anthropic/claude-3.5-haiku-20241022", - name: "Anthropic: Claude 3.5 Haiku (2024-10-22)", + "anthropic/claude-3.5-haiku": { + id: "anthropic/claude-3.5-haiku", + name: "Anthropic: Claude 3.5 Haiku", api: "openai-completions", provider: "openrouter", baseUrl: "https://openrouter.ai/api/v1", @@ -6040,9 +5950,9 @@ export const MODELS = { contextWindow: 200000, maxTokens: 8192, } satisfies Model<"openai-completions">, - "anthropic/claude-3.5-haiku": { - id: "anthropic/claude-3.5-haiku", - name: "Anthropic: Claude 3.5 Haiku", + "anthropic/claude-3.5-haiku-20241022": { + id: "anthropic/claude-3.5-haiku-20241022", + name: "Anthropic: Claude 3.5 Haiku (2024-10-22)", api: "openai-completions", provider: "openrouter", baseUrl: "https://openrouter.ai/api/v1", @@ -6074,23 +5984,6 @@ export const MODELS = { contextWindow: 200000, maxTokens: 8192, } satisfies Model<"openai-completions">, - "mistralai/ministral-8b": { - id: "mistralai/ministral-8b", - name: "Mistral: Ministral 8B", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 0.09999999999999999, - output: 0.09999999999999999, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 131072, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "mistralai/ministral-3b": { id: "mistralai/ministral-3b", name: "Mistral: Ministral 3B", @@ -6108,6 +6001,23 @@ export const MODELS = { contextWindow: 131072, maxTokens: 4096, } satisfies Model<"openai-completions">, + "mistralai/ministral-8b": { + id: "mistralai/ministral-8b", + name: "Mistral: Ministral 8B", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 0.09999999999999999, + output: 0.09999999999999999, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 131072, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "nvidia/llama-3.1-nemotron-70b-instruct": { id: "nvidia/llama-3.1-nemotron-70b-instruct", name: "NVIDIA: Llama 3.1 Nemotron 70B Instruct", @@ -6295,23 +6205,6 @@ export const MODELS = { contextWindow: 131072, maxTokens: 16384, } satisfies Model<"openai-completions">, - "meta-llama/llama-3.1-405b-instruct": { - id: "meta-llama/llama-3.1-405b-instruct", - name: "Meta: Llama 3.1 405B Instruct", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 3.5, - output: 3.5, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 130815, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "meta-llama/llama-3.1-70b-instruct": { id: "meta-llama/llama-3.1-70b-instruct", name: "Meta: Llama 3.1 70B Instruct", @@ -6329,6 +6222,23 @@ export const MODELS = { contextWindow: 131072, maxTokens: 4096, } satisfies Model<"openai-completions">, + "meta-llama/llama-3.1-405b-instruct": { + id: "meta-llama/llama-3.1-405b-instruct", + name: "Meta: Llama 3.1 405B Instruct", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 3.5, + output: 3.5, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 130815, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "mistralai/mistral-nemo": { id: "mistralai/mistral-nemo", name: "Mistral: Mistral Nemo", @@ -6465,23 +6375,6 @@ export const MODELS = { contextWindow: 128000, maxTokens: 4096, } satisfies Model<"openai-completions">, - "openai/gpt-4o-2024-05-13": { - id: "openai/gpt-4o-2024-05-13", - name: "OpenAI: GPT-4o (2024-05-13)", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text", "image"], - cost: { - input: 5, - output: 15, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 128000, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "openai/gpt-4o": { id: "openai/gpt-4o", name: "OpenAI: GPT-4o", @@ -6516,22 +6409,22 @@ export const MODELS = { contextWindow: 128000, maxTokens: 64000, } satisfies Model<"openai-completions">, - "meta-llama/llama-3-70b-instruct": { - id: "meta-llama/llama-3-70b-instruct", - name: "Meta: Llama 3 70B Instruct", + "openai/gpt-4o-2024-05-13": { + id: "openai/gpt-4o-2024-05-13", + name: "OpenAI: GPT-4o (2024-05-13)", api: "openai-completions", provider: "openrouter", baseUrl: "https://openrouter.ai/api/v1", reasoning: false, - input: ["text"], + input: ["text", "image"], cost: { - input: 0.3, - output: 0.39999999999999997, + input: 5, + output: 15, cacheRead: 0, cacheWrite: 0, }, - contextWindow: 8192, - maxTokens: 16384, + contextWindow: 128000, + maxTokens: 4096, } satisfies Model<"openai-completions">, "meta-llama/llama-3-8b-instruct": { id: "meta-llama/llama-3-8b-instruct", @@ -6550,6 +6443,23 @@ export const MODELS = { contextWindow: 8192, maxTokens: 16384, } satisfies Model<"openai-completions">, + "meta-llama/llama-3-70b-instruct": { + id: "meta-llama/llama-3-70b-instruct", + name: "Meta: Llama 3 70B Instruct", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 0.3, + output: 0.39999999999999997, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 8192, + maxTokens: 16384, + } satisfies Model<"openai-completions">, "mistralai/mixtral-8x22b-instruct": { id: "mistralai/mixtral-8x22b-instruct", name: "Mistral: Mixtral 8x22B Instruct", @@ -6754,23 +6664,6 @@ export const MODELS = { contextWindow: 8191, maxTokens: 4096, } satisfies Model<"openai-completions">, - "openai/gpt-4": { - id: "openai/gpt-4", - name: "OpenAI: GPT-4", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 30, - output: 60, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 8191, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "openai/gpt-3.5-turbo": { id: "openai/gpt-3.5-turbo", name: "OpenAI: GPT-3.5 Turbo", @@ -6788,6 +6681,23 @@ export const MODELS = { contextWindow: 16385, maxTokens: 4096, } satisfies Model<"openai-completions">, + "openai/gpt-4": { + id: "openai/gpt-4", + name: "OpenAI: GPT-4", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 30, + output: 60, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 8191, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "openrouter/auto": { id: "openrouter/auto", name: "OpenRouter: Auto Router", diff --git a/packages/ai/src/providers/anthropic.ts b/packages/ai/src/providers/anthropic.ts index 08d4e4f6..9df2568b 100644 --- a/packages/ai/src/providers/anthropic.ts +++ b/packages/ai/src/providers/anthropic.ts @@ -285,7 +285,7 @@ function createClient( const defaultHeaders = { accept: "application/json", "anthropic-dangerous-direct-browser-access": "true", - "anthropic-beta": "oauth-2025-04-20,fine-grained-tool-streaming-2025-05-14", + "anthropic-beta": "oauth-2025-04-20,fine-grained-tool-streaming-2025-05-14,interleaved-thinking-2025-05-14", ...(model.headers || {}), }; @@ -303,7 +303,7 @@ function createClient( const defaultHeaders = { accept: "application/json", "anthropic-dangerous-direct-browser-access": "true", - "anthropic-beta": "fine-grained-tool-streaming-2025-05-14", + "anthropic-beta": "fine-grained-tool-streaming-2025-05-14,interleaved-thinking-2025-05-14", ...(model.headers || {}), }; diff --git a/packages/coding-agent/CHANGELOG.md b/packages/coding-agent/CHANGELOG.md index 10c2dabd..b7959060 100644 --- a/packages/coding-agent/CHANGELOG.md +++ b/packages/coding-agent/CHANGELOG.md @@ -2,6 +2,14 @@ ## [Unreleased] +## [0.22.1] - 2025-12-15 + +_Dedicated to Peter's shoulder ([@steipete](https://twitter.com/steipete))_ + +### Changed + +- Updated `@mariozechner/pi-ai` with interleaved thinking support for Anthropic models. + ## [0.22.0] - 2025-12-15 ### Added diff --git a/packages/coding-agent/package.json b/packages/coding-agent/package.json index 87c89883..efbf467c 100644 --- a/packages/coding-agent/package.json +++ b/packages/coding-agent/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-coding-agent", - "version": "0.22.0", + "version": "0.22.1", "description": "Coding agent CLI with read, bash, edit, write tools and session management", "type": "module", "piConfig": { @@ -39,9 +39,9 @@ "prepublishOnly": "npm run clean && npm run build" }, "dependencies": { - "@mariozechner/pi-agent-core": "^0.22.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1", "chalk": "^5.5.0", "diff": "^8.0.2", "glob": "^11.0.3", diff --git a/packages/mom/package.json b/packages/mom/package.json index 10488d6b..2875d3bf 100644 --- a/packages/mom/package.json +++ b/packages/mom/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-mom", - "version": "0.22.0", + "version": "0.22.1", "description": "Slack bot that delegates messages to the pi coding agent", "type": "module", "bin": { @@ -21,9 +21,9 @@ }, "dependencies": { "@anthropic-ai/sandbox-runtime": "^0.0.16", - "@mariozechner/pi-agent-core": "^0.22.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-coding-agent": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-coding-agent": "^0.22.1", "@sinclair/typebox": "^0.34.0", "@slack/socket-mode": "^2.0.0", "@slack/web-api": "^7.0.0", diff --git a/packages/pods/package.json b/packages/pods/package.json index 56747398..0bfaf943 100644 --- a/packages/pods/package.json +++ b/packages/pods/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi", - "version": "0.22.0", + "version": "0.22.1", "description": "CLI tool for managing vLLM deployments on GPU pods", "type": "module", "bin": { @@ -34,7 +34,7 @@ "node": ">=20.0.0" }, "dependencies": { - "@mariozechner/pi-agent-core": "^0.22.0", + "@mariozechner/pi-agent-core": "^0.22.1", "chalk": "^5.5.0" }, "devDependencies": {} diff --git a/packages/proxy/package.json b/packages/proxy/package.json index a3111355..d5d1bd30 100644 --- a/packages/proxy/package.json +++ b/packages/proxy/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-proxy", - "version": "0.22.0", + "version": "0.22.1", "type": "module", "description": "CORS and authentication proxy for pi-ai", "main": "dist/index.js", diff --git a/packages/tui/package.json b/packages/tui/package.json index d79c1db6..4c00c390 100644 --- a/packages/tui/package.json +++ b/packages/tui/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-tui", - "version": "0.22.0", + "version": "0.22.1", "description": "Terminal User Interface library with differential rendering for efficient text-based applications", "type": "module", "main": "dist/index.js", diff --git a/packages/web-ui/example/package.json b/packages/web-ui/example/package.json index e4c3a35d..7b5d0435 100644 --- a/packages/web-ui/example/package.json +++ b/packages/web-ui/example/package.json @@ -1,6 +1,6 @@ { "name": "pi-web-ui-example", - "version": "1.10.0", + "version": "1.10.1", "private": true, "type": "module", "scripts": { diff --git a/packages/web-ui/package.json b/packages/web-ui/package.json index d0f6b33e..c77c326f 100644 --- a/packages/web-ui/package.json +++ b/packages/web-ui/package.json @@ -1,6 +1,6 @@ { "name": "@mariozechner/pi-web-ui", - "version": "0.22.0", + "version": "0.22.1", "description": "Reusable web UI components for AI chat interfaces powered by @mariozechner/pi-ai", "type": "module", "main": "dist/index.js", @@ -18,8 +18,8 @@ }, "dependencies": { "@lmstudio/sdk": "^1.5.0", - "@mariozechner/pi-ai": "^0.22.0", - "@mariozechner/pi-tui": "^0.22.0", + "@mariozechner/pi-ai": "^0.22.1", + "@mariozechner/pi-tui": "^0.22.1", "docx-preview": "^0.3.7", "jszip": "^3.10.1", "lucide": "^0.544.0", diff --git a/pi-mono.code-workspace b/pi-mono.code-workspace index 525d385c..8b54d2ba 100644 --- a/pi-mono.code-workspace +++ b/pi-mono.code-workspace @@ -12,6 +12,9 @@ }, { "path": "../pi-skills" + }, + { + "path": "../sitegeist" } ], "settings": {}