mirror of
https://github.com/getcompanion-ai/co-mono.git
synced 2026-04-16 04:01:56 +00:00
Fix lints
This commit is contained in:
parent
92bad8619c
commit
904fc909c9
2 changed files with 6 additions and 7 deletions
|
|
@ -1,6 +1,5 @@
|
|||
import { Button, html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit";
|
||||
import { html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit";
|
||||
import { getProviders } from "@mariozechner/pi-ai";
|
||||
import { LitElement } from "lit";
|
||||
import { customElement, state } from "lit/decorators.js";
|
||||
import "../components/CustomProviderCard.js";
|
||||
import "../components/ProviderKeyInput.js";
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ import { Ollama } from "ollama/browser";
|
|||
* @param apiKey - Optional API key (currently unused by Ollama)
|
||||
* @returns Array of discovered models
|
||||
*/
|
||||
export async function discoverOllamaModels(baseUrl: string, apiKey?: string): Promise<Model<any>[]> {
|
||||
export async function discoverOllamaModels(baseUrl: string, _apiKey?: string): Promise<Model<any>[]> {
|
||||
try {
|
||||
// Create Ollama client
|
||||
const ollama = new Ollama({ host: baseUrl });
|
||||
|
|
@ -89,7 +89,7 @@ export async function discoverLlamaCppModels(baseUrl: string, apiKey?: string):
|
|||
};
|
||||
|
||||
if (apiKey) {
|
||||
headers["Authorization"] = `Bearer ${apiKey}`;
|
||||
headers.Authorization = `Bearer ${apiKey}`;
|
||||
}
|
||||
|
||||
const response = await fetch(`${baseUrl}/v1/models`, {
|
||||
|
|
@ -151,7 +151,7 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom
|
|||
};
|
||||
|
||||
if (apiKey) {
|
||||
headers["Authorization"] = `Bearer ${apiKey}`;
|
||||
headers.Authorization = `Bearer ${apiKey}`;
|
||||
}
|
||||
|
||||
const response = await fetch(`${baseUrl}/v1/models`, {
|
||||
|
|
@ -206,11 +206,11 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom
|
|||
* @param apiKey - Optional API key (unused for LM Studio SDK)
|
||||
* @returns Array of discovered models
|
||||
*/
|
||||
export async function discoverLMStudioModels(baseUrl: string, apiKey?: string): Promise<Model<any>[]> {
|
||||
export async function discoverLMStudioModels(baseUrl: string, _apiKey?: string): Promise<Model<any>[]> {
|
||||
try {
|
||||
// Extract host and port from baseUrl
|
||||
const url = new URL(baseUrl);
|
||||
const port = url.port ? parseInt(url.port) : 1234;
|
||||
const port = url.port ? parseInt(url.port, 10) : 1234;
|
||||
|
||||
// Create LM Studio client
|
||||
const client = new LMStudioClient({ baseUrl: `ws://${url.hostname}:${port}` });
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue