From 24fcca5b7bb6e0c93e5dfcebef401a6a8d9376fe Mon Sep 17 00:00:00 2001 From: Mario Zechner Date: Wed, 3 Dec 2025 14:16:00 +0100 Subject: [PATCH] Try fixing root level compiles --- package-lock.json | 20 +-- packages/ai/src/models.generated.ts | 252 ++++++++++++++++------------ tsconfig.base.json | 22 --- 3 files changed, 153 insertions(+), 141 deletions(-) diff --git a/package-lock.json b/package-lock.json index 752d44cf..e5c816f1 100644 --- a/package-lock.json +++ b/package-lock.json @@ -6077,8 +6077,8 @@ "version": "0.12.4", "license": "MIT", "dependencies": { - "@mariozechner/pi-ai": "^0.12.3", - "@mariozechner/pi-tui": "^0.12.3" + "@mariozechner/pi-ai": "^0.12.4", + "@mariozechner/pi-tui": "^0.12.4" }, "devDependencies": { "@types/node": "^24.3.0", @@ -6152,9 +6152,9 @@ "version": "0.12.4", "license": "MIT", "dependencies": { - "@mariozechner/pi-agent-core": "^0.12.3", - "@mariozechner/pi-ai": "^0.12.3", - "@mariozechner/pi-tui": "^0.12.3", + "@mariozechner/pi-agent-core": "^0.12.4", + "@mariozechner/pi-ai": "^0.12.4", + "@mariozechner/pi-tui": "^0.12.4", "chalk": "^5.5.0", "diff": "^8.0.2", "glob": "^11.0.3" @@ -6195,8 +6195,8 @@ "license": "MIT", "dependencies": { "@anthropic-ai/sandbox-runtime": "^0.0.16", - "@mariozechner/pi-agent-core": "^0.12.3", - "@mariozechner/pi-ai": "^0.12.3", + "@mariozechner/pi-agent-core": "^0.12.4", + "@mariozechner/pi-ai": "^0.12.4", "@sinclair/typebox": "^0.34.0", "@slack/socket-mode": "^2.0.0", "@slack/web-api": "^7.0.0", @@ -6237,7 +6237,7 @@ "version": "0.12.4", "license": "MIT", "dependencies": { - "@mariozechner/pi-agent-core": "^0.12.3", + "@mariozechner/pi-agent-core": "^0.12.4", "chalk": "^5.5.0" }, "bin": { @@ -6314,8 +6314,8 @@ "license": "MIT", "dependencies": { "@lmstudio/sdk": "^1.5.0", - "@mariozechner/pi-ai": "^0.12.3", - "@mariozechner/pi-tui": "^0.12.3", + "@mariozechner/pi-ai": "^0.12.4", + "@mariozechner/pi-tui": "^0.12.4", "docx-preview": "^0.3.7", "jszip": "^3.10.1", "lucide": "^0.544.0", diff --git a/packages/ai/src/models.generated.ts b/packages/ai/src/models.generated.ts index 95108dca..55e272f9 100644 --- a/packages/ai/src/models.generated.ts +++ b/packages/ai/src/models.generated.ts @@ -1974,6 +1974,57 @@ export const MODELS = { } satisfies Model<"anthropic-messages">, }, openrouter: { + "amazon/nova-2-lite-v1:free": { + id: "amazon/nova-2-lite-v1:free", + name: "Amazon: Nova 2 Lite (free)", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: true, + input: ["text", "image"], + cost: { + input: 0, + output: 0, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 1000000, + maxTokens: 4096, + } satisfies Model<"openai-completions">, + "amazon/nova-2-lite-v1": { + id: "amazon/nova-2-lite-v1", + name: "Amazon: Nova 2 Lite", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: true, + input: ["text", "image"], + cost: { + input: 0.3, + output: 2.5, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 1000000, + maxTokens: 4096, + } satisfies Model<"openai-completions">, + "mistralai/mistral-large-2512": { + id: "mistralai/mistral-large-2512", + name: "Mistral: Mistral Large 3 2512", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text", "image"], + cost: { + input: 0.5, + output: 1.5, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 262144, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "arcee-ai/trinity-mini:free": { id: "arcee-ai/trinity-mini:free", name: "Arcee AI: Trinity Mini (free)", @@ -2017,13 +2068,13 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.28, + input: 0.27, output: 0.39999999999999997, - cacheRead: 0, + cacheRead: 0.216, cacheWrite: 0, }, contextWindow: 163840, - maxTokens: 65536, + maxTokens: 4096, } satisfies Model<"openai-completions">, "prime-intellect/intellect-3": { id: "prime-intellect/intellect-3", @@ -2093,23 +2144,6 @@ export const MODELS = { contextWindow: 200000, maxTokens: 64000, } satisfies Model<"openai-completions">, - "openrouter/bert-nebulon-alpha": { - id: "openrouter/bert-nebulon-alpha", - name: "Bert-Nebulon Alpha", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text", "image"], - cost: { - input: 0, - output: 0, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 256000, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "allenai/olmo-3-7b-instruct": { id: "allenai/olmo-3-7b-instruct", name: "AllenAI: Olmo 3 7B Instruct", @@ -2578,13 +2612,13 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.44, - output: 1.76, - cacheRead: 0, + input: 0.43, + output: 1.75, + cacheRead: 0.0799999993, cacheWrite: 0, }, - contextWindow: 204800, - maxTokens: 131072, + contextWindow: 202752, + maxTokens: 4096, } satisfies Model<"openai-completions">, "anthropic/claude-sonnet-4.5": { id: "anthropic/claude-sonnet-4.5", @@ -2680,13 +2714,13 @@ export const MODELS = { reasoning: false, input: ["text", "image"], cost: { - input: 0.21, - output: 1.9, + input: 0.19999999999999998, + output: 1.2, cacheRead: 0, cacheWrite: 0, }, - contextWindow: 131072, - maxTokens: 32768, + contextWindow: 262144, + maxTokens: 4096, } satisfies Model<"openai-completions">, "qwen/qwen3-max": { id: "qwen/qwen3-max", @@ -2739,6 +2773,23 @@ export const MODELS = { contextWindow: 400000, maxTokens: 128000, } satisfies Model<"openai-completions">, + "deepseek/deepseek-v3.1-terminus:exacto": { + id: "deepseek/deepseek-v3.1-terminus:exacto", + name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: true, + input: ["text"], + cost: { + input: 0.21, + output: 0.7899999999999999, + cacheRead: 0.16799999999999998, + cacheWrite: 0, + }, + contextWindow: 163840, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "deepseek/deepseek-v3.1-terminus": { id: "deepseek/deepseek-v3.1-terminus", name: "DeepSeek: DeepSeek V3.1 Terminus", @@ -2756,23 +2807,6 @@ export const MODELS = { contextWindow: 163840, maxTokens: 4096, } satisfies Model<"openai-completions">, - "deepseek/deepseek-v3.1-terminus:exacto": { - id: "deepseek/deepseek-v3.1-terminus:exacto", - name: "DeepSeek: DeepSeek V3.1 Terminus (exacto)", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: true, - input: ["text"], - cost: { - input: 0.216, - output: 0.7999999999999999, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 131072, - maxTokens: 65536, - } satisfies Model<"openai-completions">, "x-ai/grok-4-fast": { id: "x-ai/grok-4-fast", name: "xAI: Grok 4 Fast", @@ -3598,13 +3632,13 @@ export const MODELS = { reasoning: true, input: ["text"], cost: { - input: 0.072, - output: 0.464, + input: 0.071, + output: 0.463, cacheRead: 0, cacheWrite: 0, }, - contextWindow: 131072, - maxTokens: 16384, + contextWindow: 262144, + maxTokens: 4096, } satisfies Model<"openai-completions">, "moonshotai/kimi-k2": { id: "moonshotai/kimi-k2", @@ -4771,8 +4805,8 @@ export const MODELS = { reasoning: false, input: ["text"], cost: { - input: 0.10400000000000001, - output: 0.312, + input: 0.108, + output: 0.32, cacheRead: 0, cacheWrite: 0, }, @@ -5085,23 +5119,6 @@ export const MODELS = { contextWindow: 32768, maxTokens: 4096, } satisfies Model<"openai-completions">, - "cohere/command-r-plus-08-2024": { - id: "cohere/command-r-plus-08-2024", - name: "Cohere: Command R+ (08-2024)", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 2.5, - output: 10, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 128000, - maxTokens: 4000, - } satisfies Model<"openai-completions">, "cohere/command-r-08-2024": { id: "cohere/command-r-08-2024", name: "Cohere: Command R (08-2024)", @@ -5119,6 +5136,23 @@ export const MODELS = { contextWindow: 128000, maxTokens: 4000, } satisfies Model<"openai-completions">, + "cohere/command-r-plus-08-2024": { + id: "cohere/command-r-plus-08-2024", + name: "Cohere: Command R+ (08-2024)", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 2.5, + output: 10, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 128000, + maxTokens: 4000, + } satisfies Model<"openai-completions">, "sao10k/l3.1-euryale-70b": { id: "sao10k/l3.1-euryale-70b", name: "Sao10K: Llama 3.1 Euryale 70B v2.2", @@ -5170,23 +5204,6 @@ export const MODELS = { contextWindow: 128000, maxTokens: 16384, } satisfies Model<"openai-completions">, - "meta-llama/llama-3.1-70b-instruct": { - id: "meta-llama/llama-3.1-70b-instruct", - name: "Meta: Llama 3.1 70B Instruct", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 0.39999999999999997, - output: 0.39999999999999997, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 131072, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "meta-llama/llama-3.1-8b-instruct": { id: "meta-llama/llama-3.1-8b-instruct", name: "Meta: Llama 3.1 8B Instruct", @@ -5221,6 +5238,23 @@ export const MODELS = { contextWindow: 130815, maxTokens: 4096, } satisfies Model<"openai-completions">, + "meta-llama/llama-3.1-70b-instruct": { + id: "meta-llama/llama-3.1-70b-instruct", + name: "Meta: Llama 3.1 70B Instruct", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 0.39999999999999997, + output: 0.39999999999999997, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 131072, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "mistralai/mistral-nemo": { id: "mistralai/mistral-nemo", name: "Mistral: Mistral Nemo", @@ -5238,9 +5272,9 @@ export const MODELS = { contextWindow: 131072, maxTokens: 16384, } satisfies Model<"openai-completions">, - "openai/gpt-4o-mini": { - id: "openai/gpt-4o-mini", - name: "OpenAI: GPT-4o-mini", + "openai/gpt-4o-mini-2024-07-18": { + id: "openai/gpt-4o-mini-2024-07-18", + name: "OpenAI: GPT-4o-mini (2024-07-18)", api: "openai-completions", provider: "openrouter", baseUrl: "https://openrouter.ai/api/v1", @@ -5255,9 +5289,9 @@ export const MODELS = { contextWindow: 128000, maxTokens: 16384, } satisfies Model<"openai-completions">, - "openai/gpt-4o-mini-2024-07-18": { - id: "openai/gpt-4o-mini-2024-07-18", - name: "OpenAI: GPT-4o-mini (2024-07-18)", + "openai/gpt-4o-mini": { + id: "openai/gpt-4o-mini", + name: "OpenAI: GPT-4o-mini", api: "openai-completions", provider: "openrouter", baseUrl: "https://openrouter.ai/api/v1", @@ -5646,23 +5680,6 @@ export const MODELS = { contextWindow: 16385, maxTokens: 4096, } satisfies Model<"openai-completions">, - "openai/gpt-3.5-turbo": { - id: "openai/gpt-3.5-turbo", - name: "OpenAI: GPT-3.5 Turbo", - api: "openai-completions", - provider: "openrouter", - baseUrl: "https://openrouter.ai/api/v1", - reasoning: false, - input: ["text"], - cost: { - input: 0.5, - output: 1.5, - cacheRead: 0, - cacheWrite: 0, - }, - contextWindow: 16385, - maxTokens: 4096, - } satisfies Model<"openai-completions">, "openai/gpt-4-0314": { id: "openai/gpt-4-0314", name: "OpenAI: GPT-4 (older v0314)", @@ -5697,6 +5714,23 @@ export const MODELS = { contextWindow: 8191, maxTokens: 4096, } satisfies Model<"openai-completions">, + "openai/gpt-3.5-turbo": { + id: "openai/gpt-3.5-turbo", + name: "OpenAI: GPT-3.5 Turbo", + api: "openai-completions", + provider: "openrouter", + baseUrl: "https://openrouter.ai/api/v1", + reasoning: false, + input: ["text"], + cost: { + input: 0.5, + output: 1.5, + cacheRead: 0, + cacheWrite: 0, + }, + contextWindow: 16385, + maxTokens: 4096, + } satisfies Model<"openai-completions">, "openrouter/auto": { id: "openrouter/auto", name: "OpenRouter: Auto Router", diff --git a/tsconfig.base.json b/tsconfig.base.json index 36c027a6..bd9a33a4 100644 --- a/tsconfig.base.json +++ b/tsconfig.base.json @@ -3,28 +3,6 @@ "target": "ES2022", "module": "Node16", "lib": ["ES2022"], - "paths": { - "*": ["./*"], - "@mariozechner/pi-ai": ["./packages/ai/src/index.ts"], - "@mariozechner/pi-ai/*": ["./packages/ai/src/*"], - "@mariozechner/pi-ai/dist/*": ["./packages/ai/src/*"], - "@mariozechner/pi-agent-core": ["./packages/agent/src/index.ts"], - "@mariozechner/pi-agent-core/*": ["./packages/agent/src/*"], - "@mariozechner/pi-coding-agent": ["./packages/coding-agent/src/index.ts"], - "@mariozechner/pi-coding-agent/*": ["./packages/coding-agent/src/*"], - "@mariozechner/pi-mom": ["./packages/mom/src/index.ts"], - "@mariozechner/pi-mom/*": ["./packages/mom/src/*"], - "@mariozechner/pi": ["./packages/pods/src/index.ts"], - "@mariozechner/pi/*": ["./packages/pods/src/*"], - "@mariozechner/pi-proxy": ["./packages/proxy/src/index.ts"], - "@mariozechner/pi-proxy/*": ["./packages/proxy/src/*"], - "@mariozechner/pi-tui": ["./packages/tui/src/index.ts"], - "@mariozechner/pi-tui/*": ["./packages/tui/src/*"], - "@mariozechner/pi-web-ui": ["./packages/web-ui/src/index.ts"], - "@mariozechner/pi-web-ui/*": ["./packages/web-ui/src/*"], - "@mariozechner/pi-agent-old": ["./packages/agent-old/src/index.ts"], - "@mariozechner/pi-agent-old/*": ["./packages/agent-old/src/*"] - }, "strict": true, "esModuleInterop": true, "skipLibCheck": true,