From 904fc909c9061f3de83e535a19d211b2e11e5d97 Mon Sep 17 00:00:00 2001 From: Mario Zechner Date: Mon, 10 Nov 2025 21:55:55 +0100 Subject: [PATCH] Fix lints --- packages/web-ui/src/dialogs/ProvidersModelsTab.ts | 3 +-- packages/web-ui/src/utils/model-discovery.ts | 10 +++++----- 2 files changed, 6 insertions(+), 7 deletions(-) diff --git a/packages/web-ui/src/dialogs/ProvidersModelsTab.ts b/packages/web-ui/src/dialogs/ProvidersModelsTab.ts index ad48ebb0..2246111b 100644 --- a/packages/web-ui/src/dialogs/ProvidersModelsTab.ts +++ b/packages/web-ui/src/dialogs/ProvidersModelsTab.ts @@ -1,6 +1,5 @@ -import { Button, html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit"; +import { html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit"; import { getProviders } from "@mariozechner/pi-ai"; -import { LitElement } from "lit"; import { customElement, state } from "lit/decorators.js"; import "../components/CustomProviderCard.js"; import "../components/ProviderKeyInput.js"; diff --git a/packages/web-ui/src/utils/model-discovery.ts b/packages/web-ui/src/utils/model-discovery.ts index 805b9d5b..299aa9ac 100644 --- a/packages/web-ui/src/utils/model-discovery.ts +++ b/packages/web-ui/src/utils/model-discovery.ts @@ -8,7 +8,7 @@ import { Ollama } from "ollama/browser"; * @param apiKey - Optional API key (currently unused by Ollama) * @returns Array of discovered models */ -export async function discoverOllamaModels(baseUrl: string, apiKey?: string): Promise[]> { +export async function discoverOllamaModels(baseUrl: string, _apiKey?: string): Promise[]> { try { // Create Ollama client const ollama = new Ollama({ host: baseUrl }); @@ -89,7 +89,7 @@ export async function discoverLlamaCppModels(baseUrl: string, apiKey?: string): }; if (apiKey) { - headers["Authorization"] = `Bearer ${apiKey}`; + headers.Authorization = `Bearer ${apiKey}`; } const response = await fetch(`${baseUrl}/v1/models`, { @@ -151,7 +151,7 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom }; if (apiKey) { - headers["Authorization"] = `Bearer ${apiKey}`; + headers.Authorization = `Bearer ${apiKey}`; } const response = await fetch(`${baseUrl}/v1/models`, { @@ -206,11 +206,11 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom * @param apiKey - Optional API key (unused for LM Studio SDK) * @returns Array of discovered models */ -export async function discoverLMStudioModels(baseUrl: string, apiKey?: string): Promise[]> { +export async function discoverLMStudioModels(baseUrl: string, _apiKey?: string): Promise[]> { try { // Extract host and port from baseUrl const url = new URL(baseUrl); - const port = url.port ? parseInt(url.port) : 1234; + const port = url.port ? parseInt(url.port, 10) : 1234; // Create LM Studio client const client = new LMStudioClient({ baseUrl: `ws://${url.hostname}:${port}` });