Add ollama dependency and dialog backdrop blur

- Add ollama package to web-ui dependencies for ModelSelector
- Add backdrop blur to SettingsDialog (bg-black/50 backdrop-blur-sm)
- Update mini-lit to 0.1.9 for backdropClassName prop support
- Fix TypeScript errors in ModelSelector (ollama import, parameter types)
- Add backward compatibility methods to SessionsStore (saveSession, loadSession, getLatestSessionId)
This commit is contained in:
Mario Zechner 2025-10-08 17:49:56 +02:00
parent db34ef01bf
commit 91c1dc6475
10 changed files with 133 additions and 70 deletions

26
package-lock.json generated
View file

@ -817,9 +817,9 @@
} }
}, },
"node_modules/@mariozechner/mini-lit": { "node_modules/@mariozechner/mini-lit": {
"version": "0.1.8", "version": "0.1.9",
"resolved": "https://registry.npmjs.org/@mariozechner/mini-lit/-/mini-lit-0.1.8.tgz", "resolved": "https://registry.npmjs.org/@mariozechner/mini-lit/-/mini-lit-0.1.9.tgz",
"integrity": "sha512-7jG0ESWq7emYgyp2xj7osN+DLXGfFkepPZE0nRCoIc7idLXWAGumSz/k7cZxoNyb50iVUGdVzT3rZPr8HtGClw==", "integrity": "sha512-RT+9+76YnFowTLyjJKDk6Ll6JC9qJ0SAbttdcJcgC64B9YA03Fa/PC6UxcI0PEmE+QWUdPTEyXJ1U5UsgzQxTw==",
"dependencies": { "dependencies": {
"@preact/signals-core": "^1.12.1", "@preact/signals-core": "^1.12.1",
"class-variance-authority": "^0.7.1", "class-variance-authority": "^0.7.1",
@ -3698,6 +3698,15 @@
} }
} }
}, },
"node_modules/ollama": {
"version": "0.6.0",
"resolved": "https://registry.npmjs.org/ollama/-/ollama-0.6.0.tgz",
"integrity": "sha512-FHjdU2Ok5x2HZsxPui/MBJZ5J+HzmxoWYa/p9wk736eT+uAhS8nvIICar5YgwlG5MFNjDR6UA5F3RSKq+JseOA==",
"license": "MIT",
"dependencies": {
"whatwg-fetch": "^3.6.20"
}
},
"node_modules/once": { "node_modules/once": {
"version": "1.4.0", "version": "1.4.0",
"resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz", "resolved": "https://registry.npmjs.org/once/-/once-1.4.0.tgz",
@ -4654,6 +4663,12 @@
"integrity": "sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ==", "integrity": "sha512-2JAn3z8AR6rjK8Sm8orRC0h/bcl/DqL7tRPdGZ4I1CjdF+EaMLmYxBHyXuKL849eucPFhvBoxMsflfOb8kxaeQ==",
"license": "BSD-2-Clause" "license": "BSD-2-Clause"
}, },
"node_modules/whatwg-fetch": {
"version": "3.6.20",
"resolved": "https://registry.npmjs.org/whatwg-fetch/-/whatwg-fetch-3.6.20.tgz",
"integrity": "sha512-EqhiFU6daOA8kpjOWTL0olhVOF3i7OrFzSYiGsEMB8GcXS+RrzauAERX65xMeNWVqxA6HXH2m69Z9LaKKdisfg==",
"license": "MIT"
},
"node_modules/whatwg-url": { "node_modules/whatwg-url": {
"version": "5.0.0", "version": "5.0.0",
"resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz", "resolved": "https://registry.npmjs.org/whatwg-url/-/whatwg-url-5.0.0.tgz",
@ -5451,17 +5466,18 @@
"jszip": "^3.10.1", "jszip": "^3.10.1",
"lit": "^3.3.1", "lit": "^3.3.1",
"lucide": "^0.544.0", "lucide": "^0.544.0",
"ollama": "^0.6.0",
"pdfjs-dist": "^5.4.149", "pdfjs-dist": "^5.4.149",
"xlsx": "^0.18.5" "xlsx": "^0.18.5"
}, },
"devDependencies": { "devDependencies": {
"@mariozechner/mini-lit": "^0.1.8", "@mariozechner/mini-lit": "^0.1.9",
"@tailwindcss/cli": "^4.0.0-beta.14", "@tailwindcss/cli": "^4.0.0-beta.14",
"concurrently": "^9.2.1", "concurrently": "^9.2.1",
"typescript": "^5.7.3" "typescript": "^5.7.3"
}, },
"peerDependencies": { "peerDependencies": {
"@mariozechner/mini-lit": "^0.1.8" "@mariozechner/mini-lit": "^0.1.9"
} }
} }
} }

View file

@ -2221,13 +2221,13 @@ export const MODELS = {
reasoning: true, reasoning: true,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0, input: 0.14,
output: 0, output: 0.86,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 32768, contextWindow: 131072,
maxTokens: 32768, maxTokens: 131072,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"qwen/qwen3-235b-a22b-thinking-2507": { "qwen/qwen3-235b-a22b-thinking-2507": {
id: "qwen/qwen3-235b-a22b-thinking-2507", id: "qwen/qwen3-235b-a22b-thinking-2507",
@ -3283,23 +3283,6 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"cohere/command-r-plus-08-2024": {
id: "cohere/command-r-plus-08-2024",
name: "Cohere: Command R+ (08-2024)",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 2.5,
output: 10,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 4000,
} satisfies Model<"openai-completions">,
"cohere/command-r-08-2024": { "cohere/command-r-08-2024": {
id: "cohere/command-r-08-2024", id: "cohere/command-r-08-2024",
name: "Cohere: Command R (08-2024)", name: "Cohere: Command R (08-2024)",
@ -3317,6 +3300,23 @@ export const MODELS = {
contextWindow: 128000, contextWindow: 128000,
maxTokens: 4000, maxTokens: 4000,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"cohere/command-r-plus-08-2024": {
id: "cohere/command-r-plus-08-2024",
name: "Cohere: Command R+ (08-2024)",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 2.5,
output: 10,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 128000,
maxTokens: 4000,
} satisfies Model<"openai-completions">,
"microsoft/phi-3.5-mini-128k-instruct": { "microsoft/phi-3.5-mini-128k-instruct": {
id: "microsoft/phi-3.5-mini-128k-instruct", id: "microsoft/phi-3.5-mini-128k-instruct",
name: "Microsoft: Phi-3.5 Mini 128K Instruct", name: "Microsoft: Phi-3.5 Mini 128K Instruct",
@ -3351,22 +3351,22 @@ export const MODELS = {
contextWindow: 65000, contextWindow: 65000,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"meta-llama/llama-3.1-8b-instruct": { "meta-llama/llama-3.1-70b-instruct": {
id: "meta-llama/llama-3.1-8b-instruct", id: "meta-llama/llama-3.1-70b-instruct",
name: "Meta: Llama 3.1 8B Instruct", name: "Meta: Llama 3.1 70B Instruct",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
reasoning: false, reasoning: false,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.02, input: 0.39999999999999997,
output: 0.03, output: 0.39999999999999997,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 131072, contextWindow: 131072,
maxTokens: 16384, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"meta-llama/llama-3.1-405b-instruct": { "meta-llama/llama-3.1-405b-instruct": {
id: "meta-llama/llama-3.1-405b-instruct", id: "meta-llama/llama-3.1-405b-instruct",
@ -3385,22 +3385,22 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 16384, maxTokens: 16384,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"meta-llama/llama-3.1-70b-instruct": { "meta-llama/llama-3.1-8b-instruct": {
id: "meta-llama/llama-3.1-70b-instruct", id: "meta-llama/llama-3.1-8b-instruct",
name: "Meta: Llama 3.1 70B Instruct", name: "Meta: Llama 3.1 8B Instruct",
api: "openai-completions", api: "openai-completions",
provider: "openrouter", provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1", baseUrl: "https://openrouter.ai/api/v1",
reasoning: false, reasoning: false,
input: ["text"], input: ["text"],
cost: { cost: {
input: 0.39999999999999997, input: 0.02,
output: 0.39999999999999997, output: 0.03,
cacheRead: 0, cacheRead: 0,
cacheWrite: 0, cacheWrite: 0,
}, },
contextWindow: 131072, contextWindow: 131072,
maxTokens: 4096, maxTokens: 16384,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/mistral-nemo": { "mistralai/mistral-nemo": {
id: "mistralai/mistral-nemo", id: "mistralai/mistral-nemo",
@ -3572,23 +3572,6 @@ export const MODELS = {
contextWindow: 128000, contextWindow: 128000,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/mistral-small": {
id: "mistralai/mistral-small",
name: "Mistral Small",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.19999999999999998,
output: 0.6,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 32768,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/mistral-tiny": { "mistralai/mistral-tiny": {
id: "mistralai/mistral-tiny", id: "mistralai/mistral-tiny",
name: "Mistral Tiny", name: "Mistral Tiny",
@ -3606,6 +3589,23 @@ export const MODELS = {
contextWindow: 32768, contextWindow: 32768,
maxTokens: 4096, maxTokens: 4096,
} satisfies Model<"openai-completions">, } satisfies Model<"openai-completions">,
"mistralai/mistral-small": {
id: "mistralai/mistral-small",
name: "Mistral Small",
api: "openai-completions",
provider: "openrouter",
baseUrl: "https://openrouter.ai/api/v1",
reasoning: false,
input: ["text"],
cost: {
input: 0.19999999999999998,
output: 0.6,
cacheRead: 0,
cacheWrite: 0,
},
contextWindow: 32768,
maxTokens: 4096,
} satisfies Model<"openai-completions">,
"mistralai/mixtral-8x7b-instruct": { "mistralai/mixtral-8x7b-instruct": {
id: "mistralai/mixtral-8x7b-instruct", id: "mistralai/mixtral-8x7b-instruct",
name: "Mistral: Mixtral 8x7B Instruct", name: "Mistral: Mixtral 8x7B Instruct",

View file

@ -22,14 +22,15 @@
"jszip": "^3.10.1", "jszip": "^3.10.1",
"lit": "^3.3.1", "lit": "^3.3.1",
"lucide": "^0.544.0", "lucide": "^0.544.0",
"ollama": "^0.6.0",
"pdfjs-dist": "^5.4.149", "pdfjs-dist": "^5.4.149",
"xlsx": "^0.18.5" "xlsx": "^0.18.5"
}, },
"peerDependencies": { "peerDependencies": {
"@mariozechner/mini-lit": "^0.1.8" "@mariozechner/mini-lit": "^0.1.9"
}, },
"devDependencies": { "devDependencies": {
"@mariozechner/mini-lit": "^0.1.8", "@mariozechner/mini-lit": "^0.1.9",
"@tailwindcss/cli": "^4.0.0-beta.14", "@tailwindcss/cli": "^4.0.0-beta.14",
"concurrently": "^9.2.1", "concurrently": "^9.2.1",
"typescript": "^5.7.3" "typescript": "^5.7.3"

View file

@ -1,4 +1,4 @@
import { Button, html, icon } from "@mariozechner/mini-lit"; import { html } from "@mariozechner/mini-lit";
import type { import type {
AgentTool, AgentTool,
AssistantMessage as AssistantMessageType, AssistantMessage as AssistantMessageType,
@ -7,8 +7,7 @@ import type {
UserMessage as UserMessageType, UserMessage as UserMessageType,
} from "@mariozechner/pi-ai"; } from "@mariozechner/pi-ai";
import { LitElement, type TemplateResult } from "lit"; import { LitElement, type TemplateResult } from "lit";
import { customElement, property, state } from "lit/decorators.js"; import { customElement, property } from "lit/decorators.js";
import { Bug, Loader, Wrench } from "lucide";
import { renderTool } from "../tools/index.js"; import { renderTool } from "../tools/index.js";
import type { Attachment } from "../utils/attachment-utils.js"; import type { Attachment } from "../utils/attachment-utils.js";
import { formatUsage } from "../utils/format.js"; import { formatUsage } from "../utils/format.js";

View file

@ -101,7 +101,7 @@ export class ModelSelector extends DialogBase {
// Fetch details for each model and convert to Model format // Fetch details for each model and convert to Model format
const ollamaModelPromises: Promise<Model<any> | null>[] = models const ollamaModelPromises: Promise<Model<any> | null>[] = models
.map(async (model) => { .map(async (model: any) => {
try { try {
// Get model details // Get model details
const details = await ollama.show({ const details = await ollama.show({
@ -144,7 +144,7 @@ export class ModelSelector extends DialogBase {
return null; return null;
} }
}) })
.filter((m) => m !== null); .filter((m: any) => m !== null);
const results = await Promise.all(ollamaModelPromises); const results = await Promise.all(ollamaModelPromises);
this.ollamaModels = results.filter((m): m is Model<any> => m !== null); this.ollamaModels = results.filter((m): m is Model<any> => m !== null);

View file

@ -179,6 +179,7 @@ export class SettingsDialog extends LitElement {
}, },
width: "min(1000px, 90vw)", width: "min(1000px, 90vw)",
height: "min(800px, 90vh)", height: "min(800px, 90vh)",
backdropClassName: "bg-black/50 backdrop-blur-sm",
children: html` children: html`
${DialogContent({ ${DialogContent({
className: "h-full p-6", className: "h-full p-6",

View file

@ -17,7 +17,7 @@ export class IndexedDBStorageBackend implements StorageBackend {
request.onerror = () => reject(request.error); request.onerror = () => reject(request.error);
request.onsuccess = () => resolve(request.result); request.onsuccess = () => resolve(request.result);
request.onupgradeneeded = (event) => { request.onupgradeneeded = (_event) => {
const db = request.result; const db = request.result;
// Create object stores from config // Create object stores from config
@ -145,7 +145,7 @@ export class IndexedDBStorageBackend implements StorageBackend {
} }
async getQuotaInfo(): Promise<{ usage: number; quota: number; percent: number }> { async getQuotaInfo(): Promise<{ usage: number; quota: number; percent: number }> {
if (navigator.storage && navigator.storage.estimate) { if (navigator.storage?.estimate) {
const estimate = await navigator.storage.estimate(); const estimate = await navigator.storage.estimate();
return { return {
usage: estimate.usage || 0, usage: estimate.usage || 0,
@ -157,7 +157,7 @@ export class IndexedDBStorageBackend implements StorageBackend {
} }
async requestPersistence(): Promise<boolean> { async requestPersistence(): Promise<boolean> {
if (navigator.storage && navigator.storage.persist) { if (navigator.storage?.persist) {
return await navigator.storage.persist(); return await navigator.storage.persist();
} }
return false; return false;

View file

@ -83,4 +83,51 @@ export class SessionsStore extends Store {
async requestPersistence(): Promise<boolean> { async requestPersistence(): Promise<boolean> {
return this.getBackend().requestPersistence(); return this.getBackend().requestPersistence();
} }
// Alias methods for backward compatibility
async saveSession(id: string, state: any, metadata: SessionMetadata | undefined, title?: string): Promise<void> {
// If metadata is provided, use it; otherwise create it from state
const meta: SessionMetadata = metadata || {
id,
title: title || "",
createdAt: new Date().toISOString(),
lastModified: new Date().toISOString(),
messageCount: state.messages?.length || 0,
usage: state.usage || {
input: 0,
output: 0,
cacheRead: 0,
cacheWrite: 0,
cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0, total: 0 },
},
modelId: state.model?.id || null,
thinkingLevel: state.thinkingLevel || "off",
preview: "",
};
const data: SessionData = {
id,
title: title || meta.title,
model: state.model,
thinkingLevel: state.thinkingLevel,
messages: state.messages || [],
createdAt: meta.createdAt,
lastModified: new Date().toISOString(),
};
await this.save(data, meta);
}
async loadSession(id: string): Promise<SessionData | null> {
return this.get(id);
}
async getLatestSessionId(): Promise<string | null> {
const allMetadata = await this.getAllMetadata();
if (allMetadata.length === 0) return null;
// Sort by lastModified descending
allMetadata.sort((a, b) => b.lastModified.localeCompare(a.lastModified));
return allMetadata[0].id;
}
} }

View file

@ -1,8 +1,8 @@
import { Diff, html, icon, iconDOM, type TemplateResult } from "@mariozechner/mini-lit"; import { Diff, html, type TemplateResult } from "@mariozechner/mini-lit";
import "@mariozechner/mini-lit/dist/CodeBlock.js"; import "@mariozechner/mini-lit/dist/CodeBlock.js";
import type { ToolResultMessage } from "@mariozechner/pi-ai"; import type { ToolResultMessage } from "@mariozechner/pi-ai";
import { createRef, ref } from "lit/directives/ref.js"; import { createRef, ref } from "lit/directives/ref.js";
import { ChevronDown, ChevronRight, FileCode2, Loader } from "lucide"; import { FileCode2 } from "lucide";
import "../../components/ConsoleBlock.js"; import "../../components/ConsoleBlock.js";
import { i18n } from "../../utils/i18n.js"; import { i18n } from "../../utils/i18n.js";
import { renderCollapsibleHeader, renderHeader } from "../renderer-registry.js"; import { renderCollapsibleHeader, renderHeader } from "../renderer-registry.js";
@ -232,7 +232,6 @@ export class ArtifactsToolRenderer implements ToolRenderer<ArtifactsParams, unde
</div> </div>
`; `;
case "delete":
default: default:
return html` return html`
<div> <div>

View file

@ -1,5 +1,6 @@
import { Button, icon } from "@mariozechner/mini-lit"; import { Button, icon } from "@mariozechner/mini-lit";
import { type AgentTool, type Message, StringEnum, type ToolCall, type ToolResultMessage } from "@mariozechner/pi-ai"; import "@mariozechner/mini-lit/dist/MarkdownBlock.js";
import { type AgentTool, type Message, StringEnum, type ToolCall } from "@mariozechner/pi-ai";
import { type Static, Type } from "@sinclair/typebox"; import { type Static, Type } from "@sinclair/typebox";
import { html, LitElement, type TemplateResult } from "lit"; import { html, LitElement, type TemplateResult } from "lit";
import { customElement, property, state } from "lit/decorators.js"; import { customElement, property, state } from "lit/decorators.js";
@ -12,7 +13,6 @@ import { HtmlArtifact } from "./HtmlArtifact.js";
import { MarkdownArtifact } from "./MarkdownArtifact.js"; import { MarkdownArtifact } from "./MarkdownArtifact.js";
import { SvgArtifact } from "./SvgArtifact.js"; import { SvgArtifact } from "./SvgArtifact.js";
import { TextArtifact } from "./TextArtifact.js"; import { TextArtifact } from "./TextArtifact.js";
import "@mariozechner/mini-lit/dist/MarkdownBlock.js";
// Simple artifact model // Simple artifact model
export interface Artifact { export interface Artifact {