mirror of
https://github.com/getcompanion-ai/co-mono.git
synced 2026-04-21 12:00:15 +00:00
Fix lints
This commit is contained in:
parent
92bad8619c
commit
904fc909c9
2 changed files with 6 additions and 7 deletions
|
|
@ -1,6 +1,5 @@
|
||||||
import { Button, html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit";
|
import { html, i18n, Select, type TemplateResult } from "@mariozechner/mini-lit";
|
||||||
import { getProviders } from "@mariozechner/pi-ai";
|
import { getProviders } from "@mariozechner/pi-ai";
|
||||||
import { LitElement } from "lit";
|
|
||||||
import { customElement, state } from "lit/decorators.js";
|
import { customElement, state } from "lit/decorators.js";
|
||||||
import "../components/CustomProviderCard.js";
|
import "../components/CustomProviderCard.js";
|
||||||
import "../components/ProviderKeyInput.js";
|
import "../components/ProviderKeyInput.js";
|
||||||
|
|
|
||||||
|
|
@ -8,7 +8,7 @@ import { Ollama } from "ollama/browser";
|
||||||
* @param apiKey - Optional API key (currently unused by Ollama)
|
* @param apiKey - Optional API key (currently unused by Ollama)
|
||||||
* @returns Array of discovered models
|
* @returns Array of discovered models
|
||||||
*/
|
*/
|
||||||
export async function discoverOllamaModels(baseUrl: string, apiKey?: string): Promise<Model<any>[]> {
|
export async function discoverOllamaModels(baseUrl: string, _apiKey?: string): Promise<Model<any>[]> {
|
||||||
try {
|
try {
|
||||||
// Create Ollama client
|
// Create Ollama client
|
||||||
const ollama = new Ollama({ host: baseUrl });
|
const ollama = new Ollama({ host: baseUrl });
|
||||||
|
|
@ -89,7 +89,7 @@ export async function discoverLlamaCppModels(baseUrl: string, apiKey?: string):
|
||||||
};
|
};
|
||||||
|
|
||||||
if (apiKey) {
|
if (apiKey) {
|
||||||
headers["Authorization"] = `Bearer ${apiKey}`;
|
headers.Authorization = `Bearer ${apiKey}`;
|
||||||
}
|
}
|
||||||
|
|
||||||
const response = await fetch(`${baseUrl}/v1/models`, {
|
const response = await fetch(`${baseUrl}/v1/models`, {
|
||||||
|
|
@ -151,7 +151,7 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom
|
||||||
};
|
};
|
||||||
|
|
||||||
if (apiKey) {
|
if (apiKey) {
|
||||||
headers["Authorization"] = `Bearer ${apiKey}`;
|
headers.Authorization = `Bearer ${apiKey}`;
|
||||||
}
|
}
|
||||||
|
|
||||||
const response = await fetch(`${baseUrl}/v1/models`, {
|
const response = await fetch(`${baseUrl}/v1/models`, {
|
||||||
|
|
@ -206,11 +206,11 @@ export async function discoverVLLMModels(baseUrl: string, apiKey?: string): Prom
|
||||||
* @param apiKey - Optional API key (unused for LM Studio SDK)
|
* @param apiKey - Optional API key (unused for LM Studio SDK)
|
||||||
* @returns Array of discovered models
|
* @returns Array of discovered models
|
||||||
*/
|
*/
|
||||||
export async function discoverLMStudioModels(baseUrl: string, apiKey?: string): Promise<Model<any>[]> {
|
export async function discoverLMStudioModels(baseUrl: string, _apiKey?: string): Promise<Model<any>[]> {
|
||||||
try {
|
try {
|
||||||
// Extract host and port from baseUrl
|
// Extract host and port from baseUrl
|
||||||
const url = new URL(baseUrl);
|
const url = new URL(baseUrl);
|
||||||
const port = url.port ? parseInt(url.port) : 1234;
|
const port = url.port ? parseInt(url.port, 10) : 1234;
|
||||||
|
|
||||||
// Create LM Studio client
|
// Create LM Studio client
|
||||||
const client = new LMStudioClient({ baseUrl: `ws://${url.hostname}:${port}` });
|
const client = new LMStudioClient({ baseUrl: `ws://${url.hostname}:${port}` });
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue