mirror of
https://github.com/harivansh-afk/sandbox-agent.git
synced 2026-04-17 07:03:31 +00:00
merge: resolve conflict with origin/main in client.ts
Keep both session config utility functions (effort-level) and process log SSE / WebSocket helpers (main). Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
commit
d9082854d6
25 changed files with 5620 additions and 207 deletions
7
.github/workflows/ci.yaml
vendored
7
.github/workflows/ci.yaml
vendored
|
|
@ -14,15 +14,16 @@ jobs:
|
||||||
- uses: dtolnay/rust-toolchain@stable
|
- uses: dtolnay/rust-toolchain@stable
|
||||||
with:
|
with:
|
||||||
components: rustfmt, clippy
|
components: rustfmt, clippy
|
||||||
- uses: Swatinem/rust-cache@main
|
- uses: Swatinem/rust-cache@v2
|
||||||
- uses: pnpm/action-setup@v4
|
- uses: pnpm/action-setup@v4
|
||||||
- uses: actions/setup-node@v4
|
- uses: actions/setup-node@v4
|
||||||
with:
|
with:
|
||||||
node-version: 20
|
node-version: 20
|
||||||
cache: pnpm
|
cache: pnpm
|
||||||
- run: pnpm install
|
- run: pnpm install
|
||||||
|
- run: npm install -g tsx
|
||||||
- name: Run checks
|
- name: Run checks
|
||||||
run: ./scripts/release/main.ts --version 0.0.0 --check
|
run: ./scripts/release/main.ts --version 0.0.0 --only-steps run-ci-checks
|
||||||
- name: Run ACP v1 server tests
|
- name: Run ACP v1 server tests
|
||||||
run: |
|
run: |
|
||||||
cargo test -p sandbox-agent-agent-management
|
cargo test -p sandbox-agent-agent-management
|
||||||
|
|
@ -31,5 +32,3 @@ jobs:
|
||||||
cargo test -p sandbox-agent --lib
|
cargo test -p sandbox-agent --lib
|
||||||
- name: Run SDK tests
|
- name: Run SDK tests
|
||||||
run: pnpm --dir sdks/typescript test
|
run: pnpm --dir sdks/typescript test
|
||||||
- name: Run Inspector browser E2E
|
|
||||||
run: pnpm --filter @sandbox-agent/inspector test:agent-browser
|
|
||||||
|
|
|
||||||
|
|
@ -32,7 +32,7 @@ schemars = "0.8"
|
||||||
utoipa = { version = "4.2", features = ["axum_extras"] }
|
utoipa = { version = "4.2", features = ["axum_extras"] }
|
||||||
|
|
||||||
# Web framework
|
# Web framework
|
||||||
axum = "0.7"
|
axum = { version = "0.7", features = ["ws"] }
|
||||||
tower = { version = "0.5", features = ["util"] }
|
tower = { version = "0.5", features = ["util"] }
|
||||||
tower-http = { version = "0.5", features = ["cors", "trace"] }
|
tower-http = { version = "0.5", features = ["cors", "trace"] }
|
||||||
|
|
||||||
|
|
|
||||||
1172
docs/openapi.json
1172
docs/openapi.json
File diff suppressed because it is too large
Load diff
380
pnpm-lock.yaml
generated
380
pnpm-lock.yaml
generated
File diff suppressed because it is too large
Load diff
|
|
@ -74,6 +74,10 @@ describe("AcpHttpClient integration", () => {
|
||||||
timeoutMs: 30000,
|
timeoutMs: 30000,
|
||||||
env: {
|
env: {
|
||||||
XDG_DATA_HOME: dataHome,
|
XDG_DATA_HOME: dataHome,
|
||||||
|
HOME: dataHome,
|
||||||
|
USERPROFILE: dataHome,
|
||||||
|
APPDATA: join(dataHome, "AppData", "Roaming"),
|
||||||
|
LOCALAPPDATA: join(dataHome, "AppData", "Local"),
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
baseUrl = handle.baseUrl;
|
baseUrl = handle.baseUrl;
|
||||||
|
|
|
||||||
|
|
@ -60,6 +60,10 @@ describe("IndexedDB persistence end-to-end", () => {
|
||||||
timeoutMs: 30000,
|
timeoutMs: 30000,
|
||||||
env: {
|
env: {
|
||||||
XDG_DATA_HOME: dataHome,
|
XDG_DATA_HOME: dataHome,
|
||||||
|
HOME: dataHome,
|
||||||
|
USERPROFILE: dataHome,
|
||||||
|
APPDATA: join(dataHome, "AppData", "Roaming"),
|
||||||
|
LOCALAPPDATA: join(dataHome, "AppData", "Local"),
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
baseUrl = handle.baseUrl;
|
baseUrl = handle.baseUrl;
|
||||||
|
|
|
||||||
|
|
@ -64,6 +64,10 @@ describe("Postgres persistence driver", () => {
|
||||||
timeoutMs: 30000,
|
timeoutMs: 30000,
|
||||||
env: {
|
env: {
|
||||||
XDG_DATA_HOME: dataHome,
|
XDG_DATA_HOME: dataHome,
|
||||||
|
HOME: dataHome,
|
||||||
|
USERPROFILE: dataHome,
|
||||||
|
APPDATA: join(dataHome, "AppData", "Roaming"),
|
||||||
|
LOCALAPPDATA: join(dataHome, "AppData", "Local"),
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
baseUrl = handle.baseUrl;
|
baseUrl = handle.baseUrl;
|
||||||
|
|
|
||||||
|
|
@ -55,6 +55,10 @@ describe("SQLite persistence driver", () => {
|
||||||
timeoutMs: 30000,
|
timeoutMs: 30000,
|
||||||
env: {
|
env: {
|
||||||
XDG_DATA_HOME: dataHome,
|
XDG_DATA_HOME: dataHome,
|
||||||
|
HOME: dataHome,
|
||||||
|
USERPROFILE: dataHome,
|
||||||
|
APPDATA: join(dataHome, "AppData", "Roaming"),
|
||||||
|
LOCALAPPDATA: join(dataHome, "AppData", "Local"),
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
baseUrl = handle.baseUrl;
|
baseUrl = handle.baseUrl;
|
||||||
|
|
|
||||||
|
|
@ -17,8 +17,8 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"acp-http-client": "workspace:*",
|
"@sandbox-agent/cli-shared": "workspace:*",
|
||||||
"@sandbox-agent/cli-shared": "workspace:*"
|
"acp-http-client": "workspace:*"
|
||||||
},
|
},
|
||||||
"files": [
|
"files": [
|
||||||
"dist"
|
"dist"
|
||||||
|
|
@ -34,10 +34,12 @@
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@types/node": "^22.0.0",
|
"@types/node": "^22.0.0",
|
||||||
|
"@types/ws": "^8.18.1",
|
||||||
"openapi-typescript": "^6.7.0",
|
"openapi-typescript": "^6.7.0",
|
||||||
"tsup": "^8.0.0",
|
"tsup": "^8.0.0",
|
||||||
"typescript": "^5.7.0",
|
"typescript": "^5.7.0",
|
||||||
"vitest": "^3.0.0"
|
"vitest": "^3.0.0",
|
||||||
|
"ws": "^8.19.0"
|
||||||
},
|
},
|
||||||
"optionalDependencies": {
|
"optionalDependencies": {
|
||||||
"@sandbox-agent/cli": "workspace:*"
|
"@sandbox-agent/cli": "workspace:*"
|
||||||
|
|
|
||||||
|
|
@ -44,6 +44,20 @@ import {
|
||||||
type McpConfigQuery,
|
type McpConfigQuery,
|
||||||
type McpServerConfig,
|
type McpServerConfig,
|
||||||
type ProblemDetails,
|
type ProblemDetails,
|
||||||
|
type ProcessConfig,
|
||||||
|
type ProcessCreateRequest,
|
||||||
|
type ProcessInfo,
|
||||||
|
type ProcessInputRequest,
|
||||||
|
type ProcessInputResponse,
|
||||||
|
type ProcessListResponse,
|
||||||
|
type ProcessLogEntry,
|
||||||
|
type ProcessLogsQuery,
|
||||||
|
type ProcessLogsResponse,
|
||||||
|
type ProcessRunRequest,
|
||||||
|
type ProcessRunResponse,
|
||||||
|
type ProcessSignalQuery,
|
||||||
|
type ProcessTerminalResizeRequest,
|
||||||
|
type ProcessTerminalResizeResponse,
|
||||||
type SessionEvent,
|
type SessionEvent,
|
||||||
type SessionPersistDriver,
|
type SessionPersistDriver,
|
||||||
type SessionRecord,
|
type SessionRecord,
|
||||||
|
|
@ -112,6 +126,27 @@ export interface SessionSendOptions {
|
||||||
}
|
}
|
||||||
|
|
||||||
export type SessionEventListener = (event: SessionEvent) => void;
|
export type SessionEventListener = (event: SessionEvent) => void;
|
||||||
|
export type ProcessLogListener = (entry: ProcessLogEntry) => void;
|
||||||
|
export type ProcessLogFollowQuery = Omit<ProcessLogsQuery, "follow">;
|
||||||
|
|
||||||
|
export interface AgentQueryOptions {
|
||||||
|
config?: boolean;
|
||||||
|
noCache?: boolean;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ProcessLogSubscription {
|
||||||
|
close(): void;
|
||||||
|
closed: Promise<void>;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ProcessTerminalWebSocketUrlOptions {
|
||||||
|
accessToken?: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ProcessTerminalConnectOptions extends ProcessTerminalWebSocketUrlOptions {
|
||||||
|
protocols?: string | string[];
|
||||||
|
WebSocket?: typeof WebSocket;
|
||||||
|
}
|
||||||
|
|
||||||
export class SandboxAgentError extends Error {
|
export class SandboxAgentError extends Error {
|
||||||
readonly status: number;
|
readonly status: number;
|
||||||
|
|
@ -1045,15 +1080,15 @@ export class SandboxAgent {
|
||||||
return this.requestJson("GET", `${API_PREFIX}/health`);
|
return this.requestJson("GET", `${API_PREFIX}/health`);
|
||||||
}
|
}
|
||||||
|
|
||||||
async listAgents(options?: { config?: boolean }): Promise<AgentListResponse> {
|
async listAgents(options?: AgentQueryOptions): Promise<AgentListResponse> {
|
||||||
return this.requestJson("GET", `${API_PREFIX}/agents`, {
|
return this.requestJson("GET", `${API_PREFIX}/agents`, {
|
||||||
query: options?.config ? { config: "true" } : undefined,
|
query: toAgentQuery(options),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
async getAgent(agent: string, options?: { config?: boolean }): Promise<AgentInfo> {
|
async getAgent(agent: string, options?: AgentQueryOptions): Promise<AgentInfo> {
|
||||||
return this.requestJson("GET", `${API_PREFIX}/agents/${encodeURIComponent(agent)}`, {
|
return this.requestJson("GET", `${API_PREFIX}/agents/${encodeURIComponent(agent)}`, {
|
||||||
query: options?.config ? { config: "true" } : undefined,
|
query: toAgentQuery(options),
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -1142,6 +1177,134 @@ export class SandboxAgent {
|
||||||
await this.requestRaw("DELETE", `${API_PREFIX}/config/skills`, { query });
|
await this.requestRaw("DELETE", `${API_PREFIX}/config/skills`, { query });
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async getProcessConfig(): Promise<ProcessConfig> {
|
||||||
|
return this.requestJson("GET", `${API_PREFIX}/processes/config`);
|
||||||
|
}
|
||||||
|
|
||||||
|
async setProcessConfig(config: ProcessConfig): Promise<ProcessConfig> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes/config`, {
|
||||||
|
body: config,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async createProcess(request: ProcessCreateRequest): Promise<ProcessInfo> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes`, {
|
||||||
|
body: request,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async runProcess(request: ProcessRunRequest): Promise<ProcessRunResponse> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes/run`, {
|
||||||
|
body: request,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async listProcesses(): Promise<ProcessListResponse> {
|
||||||
|
return this.requestJson("GET", `${API_PREFIX}/processes`);
|
||||||
|
}
|
||||||
|
|
||||||
|
async getProcess(id: string): Promise<ProcessInfo> {
|
||||||
|
return this.requestJson("GET", `${API_PREFIX}/processes/${encodeURIComponent(id)}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
async stopProcess(id: string, query?: ProcessSignalQuery): Promise<ProcessInfo> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes/${encodeURIComponent(id)}/stop`, {
|
||||||
|
query,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async killProcess(id: string, query?: ProcessSignalQuery): Promise<ProcessInfo> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes/${encodeURIComponent(id)}/kill`, {
|
||||||
|
query,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async deleteProcess(id: string): Promise<void> {
|
||||||
|
await this.requestRaw("DELETE", `${API_PREFIX}/processes/${encodeURIComponent(id)}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
async getProcessLogs(id: string, query: ProcessLogFollowQuery = {}): Promise<ProcessLogsResponse> {
|
||||||
|
return this.requestJson("GET", `${API_PREFIX}/processes/${encodeURIComponent(id)}/logs`, {
|
||||||
|
query,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async followProcessLogs(
|
||||||
|
id: string,
|
||||||
|
listener: ProcessLogListener,
|
||||||
|
query: ProcessLogFollowQuery = {},
|
||||||
|
): Promise<ProcessLogSubscription> {
|
||||||
|
const abortController = new AbortController();
|
||||||
|
const response = await this.requestRaw(
|
||||||
|
"GET",
|
||||||
|
`${API_PREFIX}/processes/${encodeURIComponent(id)}/logs`,
|
||||||
|
{
|
||||||
|
query: { ...query, follow: true },
|
||||||
|
accept: "text/event-stream",
|
||||||
|
signal: abortController.signal,
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
if (!response.body) {
|
||||||
|
abortController.abort();
|
||||||
|
throw new Error("SSE stream is not readable in this environment.");
|
||||||
|
}
|
||||||
|
|
||||||
|
const closed = consumeProcessLogSse(response.body, listener, abortController.signal);
|
||||||
|
|
||||||
|
return {
|
||||||
|
close: () => abortController.abort(),
|
||||||
|
closed,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
async sendProcessInput(id: string, request: ProcessInputRequest): Promise<ProcessInputResponse> {
|
||||||
|
return this.requestJson("POST", `${API_PREFIX}/processes/${encodeURIComponent(id)}/input`, {
|
||||||
|
body: request,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async resizeProcessTerminal(
|
||||||
|
id: string,
|
||||||
|
request: ProcessTerminalResizeRequest,
|
||||||
|
): Promise<ProcessTerminalResizeResponse> {
|
||||||
|
return this.requestJson(
|
||||||
|
"POST",
|
||||||
|
`${API_PREFIX}/processes/${encodeURIComponent(id)}/terminal/resize`,
|
||||||
|
{
|
||||||
|
body: request,
|
||||||
|
},
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
buildProcessTerminalWebSocketUrl(
|
||||||
|
id: string,
|
||||||
|
options: ProcessTerminalWebSocketUrlOptions = {},
|
||||||
|
): string {
|
||||||
|
return toWebSocketUrl(
|
||||||
|
this.buildUrl(`${API_PREFIX}/processes/${encodeURIComponent(id)}/terminal/ws`, {
|
||||||
|
access_token: options.accessToken ?? this.token,
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
connectProcessTerminalWebSocket(
|
||||||
|
id: string,
|
||||||
|
options: ProcessTerminalConnectOptions = {},
|
||||||
|
): WebSocket {
|
||||||
|
const WebSocketCtor = options.WebSocket ?? globalThis.WebSocket;
|
||||||
|
if (!WebSocketCtor) {
|
||||||
|
throw new Error("WebSocket API is not available; provide a WebSocket implementation.");
|
||||||
|
}
|
||||||
|
|
||||||
|
return new WebSocketCtor(
|
||||||
|
this.buildProcessTerminalWebSocketUrl(id, {
|
||||||
|
accessToken: options.accessToken,
|
||||||
|
}),
|
||||||
|
options.protocols,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
private async getLiveConnection(agent: string): Promise<LiveAcpConnection> {
|
private async getLiveConnection(agent: string): Promise<LiveAcpConnection> {
|
||||||
const existing = this.liveConnections.get(agent);
|
const existing = this.liveConnections.get(agent);
|
||||||
if (existing) {
|
if (existing) {
|
||||||
|
|
@ -1490,6 +1653,17 @@ async function autoAuthenticate(acp: AcpHttpClient, methods: AuthMethod[]): Prom
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function toAgentQuery(options: AgentQueryOptions | undefined): Record<string, QueryValue> | undefined {
|
||||||
|
if (!options) {
|
||||||
|
return undefined;
|
||||||
|
}
|
||||||
|
|
||||||
|
return {
|
||||||
|
config: options.config,
|
||||||
|
no_cache: options.noCache,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
function normalizeSessionInit(
|
function normalizeSessionInit(
|
||||||
value: Omit<NewSessionRequest, "_meta"> | undefined,
|
value: Omit<NewSessionRequest, "_meta"> | undefined,
|
||||||
): Omit<NewSessionRequest, "_meta"> {
|
): Omit<NewSessionRequest, "_meta"> {
|
||||||
|
|
@ -1791,3 +1965,93 @@ function toTitleCase(input: string): string {
|
||||||
.map((part) => part[0]!.toUpperCase() + part.slice(1))
|
.map((part) => part[0]!.toUpperCase() + part.slice(1))
|
||||||
.join("");
|
.join("");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function consumeProcessLogSse(
|
||||||
|
body: ReadableStream<Uint8Array>,
|
||||||
|
listener: ProcessLogListener,
|
||||||
|
signal: AbortSignal,
|
||||||
|
): Promise<void> {
|
||||||
|
const reader = body.getReader();
|
||||||
|
const decoder = new TextDecoder();
|
||||||
|
let buffer = "";
|
||||||
|
|
||||||
|
try {
|
||||||
|
while (!signal.aborted) {
|
||||||
|
const { done, value } = await reader.read();
|
||||||
|
if (done) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
buffer += decoder.decode(value, { stream: true }).replace(/\r\n/g, "\n");
|
||||||
|
|
||||||
|
let separatorIndex = buffer.indexOf("\n\n");
|
||||||
|
while (separatorIndex !== -1) {
|
||||||
|
const chunk = buffer.slice(0, separatorIndex);
|
||||||
|
buffer = buffer.slice(separatorIndex + 2);
|
||||||
|
|
||||||
|
const entry = parseProcessLogSseChunk(chunk);
|
||||||
|
if (entry) {
|
||||||
|
listener(entry);
|
||||||
|
}
|
||||||
|
|
||||||
|
separatorIndex = buffer.indexOf("\n\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
if (signal.aborted || isAbortError(error)) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
throw error;
|
||||||
|
} finally {
|
||||||
|
reader.releaseLock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function parseProcessLogSseChunk(chunk: string): ProcessLogEntry | null {
|
||||||
|
if (!chunk.trim()) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
let eventName = "message";
|
||||||
|
const dataLines: string[] = [];
|
||||||
|
|
||||||
|
for (const line of chunk.split("\n")) {
|
||||||
|
if (!line || line.startsWith(":")) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (line.startsWith("event:")) {
|
||||||
|
eventName = line.slice(6).trim();
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (line.startsWith("data:")) {
|
||||||
|
dataLines.push(line.slice(5).trimStart());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (eventName !== "log") {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
const data = dataLines.join("\n");
|
||||||
|
if (!data.trim()) {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
return JSON.parse(data) as ProcessLogEntry;
|
||||||
|
}
|
||||||
|
|
||||||
|
function toWebSocketUrl(url: string): string {
|
||||||
|
const parsed = new URL(url);
|
||||||
|
if (parsed.protocol === "http:") {
|
||||||
|
parsed.protocol = "ws:";
|
||||||
|
} else if (parsed.protocol === "https:") {
|
||||||
|
parsed.protocol = "wss:";
|
||||||
|
}
|
||||||
|
return parsed.toString();
|
||||||
|
}
|
||||||
|
|
||||||
|
function isAbortError(error: unknown): boolean {
|
||||||
|
return error instanceof Error && error.name === "AbortError";
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -57,6 +57,39 @@ export interface paths {
|
||||||
"/v1/health": {
|
"/v1/health": {
|
||||||
get: operations["get_v1_health"];
|
get: operations["get_v1_health"];
|
||||||
};
|
};
|
||||||
|
"/v1/processes": {
|
||||||
|
get: operations["get_v1_processes"];
|
||||||
|
post: operations["post_v1_processes"];
|
||||||
|
};
|
||||||
|
"/v1/processes/config": {
|
||||||
|
get: operations["get_v1_processes_config"];
|
||||||
|
post: operations["post_v1_processes_config"];
|
||||||
|
};
|
||||||
|
"/v1/processes/run": {
|
||||||
|
post: operations["post_v1_processes_run"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}": {
|
||||||
|
get: operations["get_v1_process"];
|
||||||
|
delete: operations["delete_v1_process"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/input": {
|
||||||
|
post: operations["post_v1_process_input"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/kill": {
|
||||||
|
post: operations["post_v1_process_kill"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/logs": {
|
||||||
|
get: operations["get_v1_process_logs"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/stop": {
|
||||||
|
post: operations["post_v1_process_stop"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/terminal/resize": {
|
||||||
|
post: operations["post_v1_process_terminal_resize"];
|
||||||
|
};
|
||||||
|
"/v1/processes/{id}/terminal/ws": {
|
||||||
|
get: operations["get_v1_process_terminal_ws"];
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
export type webhooks = Record<string, never>;
|
export type webhooks = Record<string, never>;
|
||||||
|
|
@ -230,6 +263,116 @@ export interface components {
|
||||||
type: string;
|
type: string;
|
||||||
[key: string]: unknown;
|
[key: string]: unknown;
|
||||||
};
|
};
|
||||||
|
ProcessConfig: {
|
||||||
|
/** Format: int64 */
|
||||||
|
defaultRunTimeoutMs: number;
|
||||||
|
maxConcurrentProcesses: number;
|
||||||
|
maxInputBytesPerRequest: number;
|
||||||
|
maxLogBytesPerProcess: number;
|
||||||
|
maxOutputBytes: number;
|
||||||
|
/** Format: int64 */
|
||||||
|
maxRunTimeoutMs: number;
|
||||||
|
};
|
||||||
|
ProcessCreateRequest: {
|
||||||
|
args?: string[];
|
||||||
|
command: string;
|
||||||
|
cwd?: string | null;
|
||||||
|
env?: {
|
||||||
|
[key: string]: string;
|
||||||
|
};
|
||||||
|
interactive?: boolean;
|
||||||
|
tty?: boolean;
|
||||||
|
};
|
||||||
|
ProcessInfo: {
|
||||||
|
args: string[];
|
||||||
|
command: string;
|
||||||
|
/** Format: int64 */
|
||||||
|
createdAtMs: number;
|
||||||
|
cwd?: string | null;
|
||||||
|
/** Format: int32 */
|
||||||
|
exitCode?: number | null;
|
||||||
|
/** Format: int64 */
|
||||||
|
exitedAtMs?: number | null;
|
||||||
|
id: string;
|
||||||
|
interactive: boolean;
|
||||||
|
/** Format: int32 */
|
||||||
|
pid?: number | null;
|
||||||
|
status: components["schemas"]["ProcessState"];
|
||||||
|
tty: boolean;
|
||||||
|
};
|
||||||
|
ProcessInputRequest: {
|
||||||
|
data: string;
|
||||||
|
encoding?: string | null;
|
||||||
|
};
|
||||||
|
ProcessInputResponse: {
|
||||||
|
bytesWritten: number;
|
||||||
|
};
|
||||||
|
ProcessListResponse: {
|
||||||
|
processes: components["schemas"]["ProcessInfo"][];
|
||||||
|
};
|
||||||
|
ProcessLogEntry: {
|
||||||
|
data: string;
|
||||||
|
encoding: string;
|
||||||
|
/** Format: int64 */
|
||||||
|
sequence: number;
|
||||||
|
stream: components["schemas"]["ProcessLogsStream"];
|
||||||
|
/** Format: int64 */
|
||||||
|
timestampMs: number;
|
||||||
|
};
|
||||||
|
ProcessLogsQuery: {
|
||||||
|
follow?: boolean | null;
|
||||||
|
/** Format: int64 */
|
||||||
|
since?: number | null;
|
||||||
|
stream?: components["schemas"]["ProcessLogsStream"] | null;
|
||||||
|
tail?: number | null;
|
||||||
|
};
|
||||||
|
ProcessLogsResponse: {
|
||||||
|
entries: components["schemas"]["ProcessLogEntry"][];
|
||||||
|
processId: string;
|
||||||
|
stream: components["schemas"]["ProcessLogsStream"];
|
||||||
|
};
|
||||||
|
/** @enum {string} */
|
||||||
|
ProcessLogsStream: "stdout" | "stderr" | "combined" | "pty";
|
||||||
|
ProcessRunRequest: {
|
||||||
|
args?: string[];
|
||||||
|
command: string;
|
||||||
|
cwd?: string | null;
|
||||||
|
env?: {
|
||||||
|
[key: string]: string;
|
||||||
|
};
|
||||||
|
maxOutputBytes?: number | null;
|
||||||
|
/** Format: int64 */
|
||||||
|
timeoutMs?: number | null;
|
||||||
|
};
|
||||||
|
ProcessRunResponse: {
|
||||||
|
/** Format: int64 */
|
||||||
|
durationMs: number;
|
||||||
|
/** Format: int32 */
|
||||||
|
exitCode?: number | null;
|
||||||
|
stderr: string;
|
||||||
|
stderrTruncated: boolean;
|
||||||
|
stdout: string;
|
||||||
|
stdoutTruncated: boolean;
|
||||||
|
timedOut: boolean;
|
||||||
|
};
|
||||||
|
ProcessSignalQuery: {
|
||||||
|
/** Format: int64 */
|
||||||
|
waitMs?: number | null;
|
||||||
|
};
|
||||||
|
/** @enum {string} */
|
||||||
|
ProcessState: "running" | "exited";
|
||||||
|
ProcessTerminalResizeRequest: {
|
||||||
|
/** Format: int32 */
|
||||||
|
cols: number;
|
||||||
|
/** Format: int32 */
|
||||||
|
rows: number;
|
||||||
|
};
|
||||||
|
ProcessTerminalResizeResponse: {
|
||||||
|
/** Format: int32 */
|
||||||
|
cols: number;
|
||||||
|
/** Format: int32 */
|
||||||
|
rows: number;
|
||||||
|
};
|
||||||
/** @enum {string} */
|
/** @enum {string} */
|
||||||
ServerStatus: "running" | "stopped";
|
ServerStatus: "running" | "stopped";
|
||||||
ServerStatusInfo: {
|
ServerStatusInfo: {
|
||||||
|
|
@ -748,4 +891,417 @@ export interface operations {
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
get_v1_processes: {
|
||||||
|
responses: {
|
||||||
|
/** @description List processes */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessListResponse"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_processes: {
|
||||||
|
requestBody: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessCreateRequest"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Started process */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInfo"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Invalid request */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process limit or state conflict */
|
||||||
|
409: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
get_v1_processes_config: {
|
||||||
|
responses: {
|
||||||
|
/** @description Current runtime process config */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessConfig"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_processes_config: {
|
||||||
|
requestBody: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessConfig"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Updated runtime process config */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessConfig"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Invalid config */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_processes_run: {
|
||||||
|
requestBody: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessRunRequest"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description One-off command result */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessRunResponse"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Invalid request */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
get_v1_process: {
|
||||||
|
parameters: {
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Process details */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInfo"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
delete_v1_process: {
|
||||||
|
parameters: {
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Process deleted */
|
||||||
|
204: {
|
||||||
|
content: never;
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process is still running */
|
||||||
|
409: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_process_input: {
|
||||||
|
parameters: {
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
requestBody: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInputRequest"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Input accepted */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInputResponse"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Invalid request */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process not writable */
|
||||||
|
409: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Input exceeds configured limit */
|
||||||
|
413: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_process_kill: {
|
||||||
|
parameters: {
|
||||||
|
query?: {
|
||||||
|
/** @description Wait up to N ms for process to exit */
|
||||||
|
waitMs?: number | null;
|
||||||
|
};
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Kill signal sent */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInfo"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
get_v1_process_logs: {
|
||||||
|
parameters: {
|
||||||
|
query?: {
|
||||||
|
/** @description stdout|stderr|combined|pty */
|
||||||
|
stream?: components["schemas"]["ProcessLogsStream"] | null;
|
||||||
|
/** @description Tail N entries */
|
||||||
|
tail?: number | null;
|
||||||
|
/** @description Follow via SSE */
|
||||||
|
follow?: boolean | null;
|
||||||
|
/** @description Only entries with sequence greater than this */
|
||||||
|
since?: number | null;
|
||||||
|
};
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Process logs */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessLogsResponse"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_process_stop: {
|
||||||
|
parameters: {
|
||||||
|
query?: {
|
||||||
|
/** @description Wait up to N ms for process to exit */
|
||||||
|
waitMs?: number | null;
|
||||||
|
};
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Stop signal sent */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessInfo"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
post_v1_process_terminal_resize: {
|
||||||
|
parameters: {
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
requestBody: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessTerminalResizeRequest"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description Resize accepted */
|
||||||
|
200: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProcessTerminalResizeResponse"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Invalid request */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Not a terminal process */
|
||||||
|
409: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
get_v1_process_terminal_ws: {
|
||||||
|
parameters: {
|
||||||
|
query?: {
|
||||||
|
/** @description Bearer token alternative for WS auth */
|
||||||
|
access_token?: string | null;
|
||||||
|
};
|
||||||
|
path: {
|
||||||
|
/** @description Process ID */
|
||||||
|
id: string;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
responses: {
|
||||||
|
/** @description WebSocket upgraded */
|
||||||
|
101: {
|
||||||
|
content: never;
|
||||||
|
};
|
||||||
|
/** @description Invalid websocket frame or upgrade request */
|
||||||
|
400: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Unknown process */
|
||||||
|
404: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Not a terminal process */
|
||||||
|
409: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
/** @description Process API unsupported on this platform */
|
||||||
|
501: {
|
||||||
|
content: {
|
||||||
|
"application/json": components["schemas"]["ProblemDetails"];
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -13,6 +13,12 @@ export { AcpRpcError } from "acp-http-client";
|
||||||
export { buildInspectorUrl } from "./inspector.ts";
|
export { buildInspectorUrl } from "./inspector.ts";
|
||||||
|
|
||||||
export type {
|
export type {
|
||||||
|
AgentQueryOptions,
|
||||||
|
ProcessLogFollowQuery,
|
||||||
|
ProcessLogListener,
|
||||||
|
ProcessLogSubscription,
|
||||||
|
ProcessTerminalConnectOptions,
|
||||||
|
ProcessTerminalWebSocketUrlOptions,
|
||||||
SandboxAgentConnectOptions,
|
SandboxAgentConnectOptions,
|
||||||
SandboxAgentStartOptions,
|
SandboxAgentStartOptions,
|
||||||
SessionCreateRequest,
|
SessionCreateRequest,
|
||||||
|
|
@ -32,6 +38,7 @@ export type {
|
||||||
AcpServerInfo,
|
AcpServerInfo,
|
||||||
AcpServerListResponse,
|
AcpServerListResponse,
|
||||||
AgentInfo,
|
AgentInfo,
|
||||||
|
AgentQuery,
|
||||||
AgentInstallRequest,
|
AgentInstallRequest,
|
||||||
AgentInstallResponse,
|
AgentInstallResponse,
|
||||||
AgentListResponse,
|
AgentListResponse,
|
||||||
|
|
@ -54,6 +61,27 @@ export type {
|
||||||
McpConfigQuery,
|
McpConfigQuery,
|
||||||
McpServerConfig,
|
McpServerConfig,
|
||||||
ProblemDetails,
|
ProblemDetails,
|
||||||
|
ProcessConfig,
|
||||||
|
ProcessCreateRequest,
|
||||||
|
ProcessInfo,
|
||||||
|
ProcessInputRequest,
|
||||||
|
ProcessInputResponse,
|
||||||
|
ProcessListResponse,
|
||||||
|
ProcessLogEntry,
|
||||||
|
ProcessLogsQuery,
|
||||||
|
ProcessLogsResponse,
|
||||||
|
ProcessLogsStream,
|
||||||
|
ProcessRunRequest,
|
||||||
|
ProcessRunResponse,
|
||||||
|
ProcessSignalQuery,
|
||||||
|
ProcessState,
|
||||||
|
ProcessTerminalClientFrame,
|
||||||
|
ProcessTerminalErrorFrame,
|
||||||
|
ProcessTerminalExitFrame,
|
||||||
|
ProcessTerminalReadyFrame,
|
||||||
|
ProcessTerminalResizeRequest,
|
||||||
|
ProcessTerminalResizeResponse,
|
||||||
|
ProcessTerminalServerFrame,
|
||||||
SessionEvent,
|
SessionEvent,
|
||||||
SessionPersistDriver,
|
SessionPersistDriver,
|
||||||
SessionRecord,
|
SessionRecord,
|
||||||
|
|
|
||||||
|
|
@ -11,6 +11,7 @@ export type ProblemDetails = components["schemas"]["ProblemDetails"];
|
||||||
export type HealthResponse = JsonResponse<operations["get_v1_health"], 200>;
|
export type HealthResponse = JsonResponse<operations["get_v1_health"], 200>;
|
||||||
export type AgentListResponse = JsonResponse<operations["get_v1_agents"], 200>;
|
export type AgentListResponse = JsonResponse<operations["get_v1_agents"], 200>;
|
||||||
export type AgentInfo = components["schemas"]["AgentInfo"];
|
export type AgentInfo = components["schemas"]["AgentInfo"];
|
||||||
|
export type AgentQuery = QueryParams<operations["get_v1_agents"]>;
|
||||||
export type AgentInstallRequest = JsonRequestBody<operations["post_v1_agent_install"]>;
|
export type AgentInstallRequest = JsonRequestBody<operations["post_v1_agent_install"]>;
|
||||||
export type AgentInstallResponse = JsonResponse<operations["post_v1_agent_install"], 200>;
|
export type AgentInstallResponse = JsonResponse<operations["post_v1_agent_install"], 200>;
|
||||||
|
|
||||||
|
|
@ -36,6 +37,58 @@ export type McpServerConfig = components["schemas"]["McpServerConfig"];
|
||||||
export type SkillsConfigQuery = QueryParams<operations["get_v1_config_skills"]>;
|
export type SkillsConfigQuery = QueryParams<operations["get_v1_config_skills"]>;
|
||||||
export type SkillsConfig = components["schemas"]["SkillsConfig"];
|
export type SkillsConfig = components["schemas"]["SkillsConfig"];
|
||||||
|
|
||||||
|
export type ProcessConfig = JsonResponse<operations["get_v1_processes_config"], 200>;
|
||||||
|
export type ProcessCreateRequest = JsonRequestBody<operations["post_v1_processes"]>;
|
||||||
|
export type ProcessInfo = components["schemas"]["ProcessInfo"];
|
||||||
|
export type ProcessInputRequest = JsonRequestBody<operations["post_v1_process_input"]>;
|
||||||
|
export type ProcessInputResponse = JsonResponse<operations["post_v1_process_input"], 200>;
|
||||||
|
export type ProcessListResponse = JsonResponse<operations["get_v1_processes"], 200>;
|
||||||
|
export type ProcessLogEntry = components["schemas"]["ProcessLogEntry"];
|
||||||
|
export type ProcessLogsQuery = QueryParams<operations["get_v1_process_logs"]>;
|
||||||
|
export type ProcessLogsResponse = JsonResponse<operations["get_v1_process_logs"], 200>;
|
||||||
|
export type ProcessLogsStream = components["schemas"]["ProcessLogsStream"];
|
||||||
|
export type ProcessRunRequest = JsonRequestBody<operations["post_v1_processes_run"]>;
|
||||||
|
export type ProcessRunResponse = JsonResponse<operations["post_v1_processes_run"], 200>;
|
||||||
|
export type ProcessSignalQuery = QueryParams<operations["post_v1_process_stop"]>;
|
||||||
|
export type ProcessState = components["schemas"]["ProcessState"];
|
||||||
|
export type ProcessTerminalResizeRequest = JsonRequestBody<operations["post_v1_process_terminal_resize"]>;
|
||||||
|
export type ProcessTerminalResizeResponse = JsonResponse<operations["post_v1_process_terminal_resize"], 200>;
|
||||||
|
|
||||||
|
export type ProcessTerminalClientFrame =
|
||||||
|
| {
|
||||||
|
type: "input";
|
||||||
|
data: string;
|
||||||
|
encoding?: string;
|
||||||
|
}
|
||||||
|
| {
|
||||||
|
type: "resize";
|
||||||
|
cols: number;
|
||||||
|
rows: number;
|
||||||
|
}
|
||||||
|
| {
|
||||||
|
type: "close";
|
||||||
|
};
|
||||||
|
|
||||||
|
export interface ProcessTerminalReadyFrame {
|
||||||
|
type: "ready";
|
||||||
|
processId: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ProcessTerminalExitFrame {
|
||||||
|
type: "exit";
|
||||||
|
exitCode?: number | null;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface ProcessTerminalErrorFrame {
|
||||||
|
type: "error";
|
||||||
|
message: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
export type ProcessTerminalServerFrame =
|
||||||
|
| ProcessTerminalReadyFrame
|
||||||
|
| ProcessTerminalExitFrame
|
||||||
|
| ProcessTerminalErrorFrame;
|
||||||
|
|
||||||
export interface SessionRecord {
|
export interface SessionRecord {
|
||||||
id: string;
|
id: string;
|
||||||
agent: string;
|
agent: string;
|
||||||
|
|
|
||||||
|
|
@ -1,18 +1,29 @@
|
||||||
import { chmodSync, mkdirSync, writeFileSync } from "node:fs";
|
import { chmodSync, mkdirSync, writeFileSync } from "node:fs";
|
||||||
import { join } from "node:path";
|
import { join } from "node:path";
|
||||||
|
|
||||||
export function prepareMockAgentDataHome(dataHome: string): void {
|
function candidateInstallDirs(dataHome: string): string[] {
|
||||||
const installDir = join(dataHome, "sandbox-agent", "bin");
|
const dirs = [join(dataHome, "sandbox-agent", "bin")];
|
||||||
const processDir = join(installDir, "agent_processes");
|
if (process.platform === "darwin") {
|
||||||
mkdirSync(processDir, { recursive: true });
|
dirs.push(join(dataHome, "Library", "Application Support", "sandbox-agent", "bin"));
|
||||||
|
} else if (process.platform === "win32") {
|
||||||
|
dirs.push(join(dataHome, "AppData", "Roaming", "sandbox-agent", "bin"));
|
||||||
|
}
|
||||||
|
return dirs;
|
||||||
|
}
|
||||||
|
|
||||||
const runner = process.platform === "win32"
|
export function prepareMockAgentDataHome(dataHome: string): Record<string, string> {
|
||||||
? join(processDir, "mock-acp.cmd")
|
const runtimeEnv: Record<string, string> = {};
|
||||||
: join(processDir, "mock-acp");
|
if (process.platform === "darwin") {
|
||||||
|
runtimeEnv.HOME = dataHome;
|
||||||
const scriptFile = process.platform === "win32"
|
runtimeEnv.XDG_DATA_HOME = join(dataHome, ".local", "share");
|
||||||
? join(processDir, "mock-acp.js")
|
} else if (process.platform === "win32") {
|
||||||
: runner;
|
runtimeEnv.USERPROFILE = dataHome;
|
||||||
|
runtimeEnv.APPDATA = join(dataHome, "AppData", "Roaming");
|
||||||
|
runtimeEnv.LOCALAPPDATA = join(dataHome, "AppData", "Local");
|
||||||
|
} else {
|
||||||
|
runtimeEnv.HOME = dataHome;
|
||||||
|
runtimeEnv.XDG_DATA_HOME = dataHome;
|
||||||
|
}
|
||||||
|
|
||||||
const nodeScript = String.raw`#!/usr/bin/env node
|
const nodeScript = String.raw`#!/usr/bin/env node
|
||||||
const { createInterface } = require("node:readline");
|
const { createInterface } = require("node:readline");
|
||||||
|
|
@ -127,14 +138,29 @@ rl.on("line", (line) => {
|
||||||
});
|
});
|
||||||
`;
|
`;
|
||||||
|
|
||||||
writeFileSync(scriptFile, nodeScript);
|
for (const installDir of candidateInstallDirs(dataHome)) {
|
||||||
|
const processDir = join(installDir, "agent_processes");
|
||||||
|
mkdirSync(processDir, { recursive: true });
|
||||||
|
|
||||||
if (process.platform === "win32") {
|
const runner = process.platform === "win32"
|
||||||
writeFileSync(runner, `@echo off\r\nnode "${scriptFile}" %*\r\n`);
|
? join(processDir, "mock-acp.cmd")
|
||||||
|
: join(processDir, "mock-acp");
|
||||||
|
|
||||||
|
const scriptFile = process.platform === "win32"
|
||||||
|
? join(processDir, "mock-acp.js")
|
||||||
|
: runner;
|
||||||
|
|
||||||
|
writeFileSync(scriptFile, nodeScript);
|
||||||
|
|
||||||
|
if (process.platform === "win32") {
|
||||||
|
writeFileSync(runner, `@echo off\r\nnode "${scriptFile}" %*\r\n`);
|
||||||
|
}
|
||||||
|
|
||||||
|
chmodSync(scriptFile, 0o755);
|
||||||
|
if (process.platform === "win32") {
|
||||||
|
chmodSync(runner, 0o755);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
chmodSync(scriptFile, 0o755);
|
return runtimeEnv;
|
||||||
if (process.platform === "win32") {
|
|
||||||
chmodSync(runner, 0o755);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -12,6 +12,7 @@ import {
|
||||||
} from "../src/index.ts";
|
} from "../src/index.ts";
|
||||||
import { spawnSandboxAgent, isNodeRuntime, type SandboxAgentSpawnHandle } from "../src/spawn.ts";
|
import { spawnSandboxAgent, isNodeRuntime, type SandboxAgentSpawnHandle } from "../src/spawn.ts";
|
||||||
import { prepareMockAgentDataHome } from "./helpers/mock-agent.ts";
|
import { prepareMockAgentDataHome } from "./helpers/mock-agent.ts";
|
||||||
|
import WebSocket from "ws";
|
||||||
|
|
||||||
const __dirname = dirname(fileURLToPath(import.meta.url));
|
const __dirname = dirname(fileURLToPath(import.meta.url));
|
||||||
|
|
||||||
|
|
@ -64,6 +65,107 @@ async function waitFor<T>(
|
||||||
throw new Error("timed out waiting for condition");
|
throw new Error("timed out waiting for condition");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function waitForAsync<T>(
|
||||||
|
fn: () => Promise<T | undefined | null>,
|
||||||
|
timeoutMs = 6000,
|
||||||
|
stepMs = 30,
|
||||||
|
): Promise<T> {
|
||||||
|
const started = Date.now();
|
||||||
|
while (Date.now() - started < timeoutMs) {
|
||||||
|
const value = await fn();
|
||||||
|
if (value !== undefined && value !== null) {
|
||||||
|
return value;
|
||||||
|
}
|
||||||
|
await sleep(stepMs);
|
||||||
|
}
|
||||||
|
throw new Error("timed out waiting for condition");
|
||||||
|
}
|
||||||
|
|
||||||
|
function buildTarArchive(entries: Array<{ name: string; content: string }>): Uint8Array {
|
||||||
|
const blocks: Buffer[] = [];
|
||||||
|
|
||||||
|
for (const entry of entries) {
|
||||||
|
const content = Buffer.from(entry.content, "utf8");
|
||||||
|
const header = Buffer.alloc(512, 0);
|
||||||
|
|
||||||
|
writeTarString(header, 0, 100, entry.name);
|
||||||
|
writeTarOctal(header, 100, 8, 0o644);
|
||||||
|
writeTarOctal(header, 108, 8, 0);
|
||||||
|
writeTarOctal(header, 116, 8, 0);
|
||||||
|
writeTarOctal(header, 124, 12, content.length);
|
||||||
|
writeTarOctal(header, 136, 12, Math.floor(Date.now() / 1000));
|
||||||
|
header.fill(0x20, 148, 156);
|
||||||
|
header[156] = "0".charCodeAt(0);
|
||||||
|
writeTarString(header, 257, 6, "ustar");
|
||||||
|
writeTarString(header, 263, 2, "00");
|
||||||
|
|
||||||
|
let checksum = 0;
|
||||||
|
for (const byte of header) {
|
||||||
|
checksum += byte;
|
||||||
|
}
|
||||||
|
writeTarChecksum(header, checksum);
|
||||||
|
|
||||||
|
blocks.push(header);
|
||||||
|
blocks.push(content);
|
||||||
|
|
||||||
|
const remainder = content.length % 512;
|
||||||
|
if (remainder !== 0) {
|
||||||
|
blocks.push(Buffer.alloc(512 - remainder, 0));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
blocks.push(Buffer.alloc(1024, 0));
|
||||||
|
return Buffer.concat(blocks);
|
||||||
|
}
|
||||||
|
|
||||||
|
function writeTarString(buffer: Buffer, offset: number, length: number, value: string): void {
|
||||||
|
const bytes = Buffer.from(value, "utf8");
|
||||||
|
bytes.copy(buffer, offset, 0, Math.min(bytes.length, length));
|
||||||
|
}
|
||||||
|
|
||||||
|
function writeTarOctal(buffer: Buffer, offset: number, length: number, value: number): void {
|
||||||
|
const rendered = value.toString(8).padStart(length - 1, "0");
|
||||||
|
writeTarString(buffer, offset, length, rendered);
|
||||||
|
buffer[offset + length - 1] = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
function writeTarChecksum(buffer: Buffer, checksum: number): void {
|
||||||
|
const rendered = checksum.toString(8).padStart(6, "0");
|
||||||
|
writeTarString(buffer, 148, 6, rendered);
|
||||||
|
buffer[154] = 0;
|
||||||
|
buffer[155] = 0x20;
|
||||||
|
}
|
||||||
|
|
||||||
|
function decodeSocketPayload(data: unknown): string {
|
||||||
|
if (typeof data === "string") {
|
||||||
|
return data;
|
||||||
|
}
|
||||||
|
if (data instanceof ArrayBuffer) {
|
||||||
|
return Buffer.from(data).toString("utf8");
|
||||||
|
}
|
||||||
|
if (ArrayBuffer.isView(data)) {
|
||||||
|
return Buffer.from(data.buffer, data.byteOffset, data.byteLength).toString("utf8");
|
||||||
|
}
|
||||||
|
if (typeof Blob !== "undefined" && data instanceof Blob) {
|
||||||
|
throw new Error("Blob socket payloads are not supported in this test");
|
||||||
|
}
|
||||||
|
throw new Error(`Unsupported socket payload type: ${typeof data}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
function decodeProcessLogData(data: string, encoding: string): string {
|
||||||
|
if (encoding === "base64") {
|
||||||
|
return Buffer.from(data, "base64").toString("utf8");
|
||||||
|
}
|
||||||
|
return data;
|
||||||
|
}
|
||||||
|
|
||||||
|
function nodeCommand(source: string): { command: string; args: string[] } {
|
||||||
|
return {
|
||||||
|
command: process.execPath,
|
||||||
|
args: ["-e", source],
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
describe("Integration: TypeScript SDK flat session API", () => {
|
describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
let handle: SandboxAgentSpawnHandle;
|
let handle: SandboxAgentSpawnHandle;
|
||||||
let baseUrl: string;
|
let baseUrl: string;
|
||||||
|
|
@ -72,15 +174,13 @@ describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
dataHome = mkdtempSync(join(tmpdir(), "sdk-integration-"));
|
dataHome = mkdtempSync(join(tmpdir(), "sdk-integration-"));
|
||||||
prepareMockAgentDataHome(dataHome);
|
const agentEnv = prepareMockAgentDataHome(dataHome);
|
||||||
|
|
||||||
handle = await spawnSandboxAgent({
|
handle = await spawnSandboxAgent({
|
||||||
enabled: true,
|
enabled: true,
|
||||||
log: "silent",
|
log: "silent",
|
||||||
timeoutMs: 30000,
|
timeoutMs: 30000,
|
||||||
env: {
|
env: agentEnv,
|
||||||
XDG_DATA_HOME: dataHome,
|
|
||||||
},
|
|
||||||
});
|
});
|
||||||
baseUrl = handle.baseUrl;
|
baseUrl = handle.baseUrl;
|
||||||
token = handle.token;
|
token = handle.token;
|
||||||
|
|
@ -122,6 +222,9 @@ describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
const fetched = await sdk.getSession(session.id);
|
const fetched = await sdk.getSession(session.id);
|
||||||
expect(fetched?.agent).toBe("mock");
|
expect(fetched?.agent).toBe("mock");
|
||||||
|
|
||||||
|
const acpServers = await sdk.listAcpServers();
|
||||||
|
expect(acpServers.servers.some((server) => server.agent === "mock")).toBe(true);
|
||||||
|
|
||||||
const events = await sdk.getEvents({ sessionId: session.id, limit: 100 });
|
const events = await sdk.getEvents({ sessionId: session.id, limit: 100 });
|
||||||
expect(events.items.length).toBeGreaterThan(0);
|
expect(events.items.length).toBeGreaterThan(0);
|
||||||
expect(events.items.some((event) => event.sender === "client")).toBe(true);
|
expect(events.items.some((event) => event.sender === "client")).toBe(true);
|
||||||
|
|
@ -137,6 +240,64 @@ describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
await sdk.dispose();
|
await sdk.dispose();
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it("covers agent query flags and filesystem HTTP helpers", async () => {
|
||||||
|
const sdk = await SandboxAgent.connect({
|
||||||
|
baseUrl,
|
||||||
|
token,
|
||||||
|
});
|
||||||
|
|
||||||
|
const directory = mkdtempSync(join(tmpdir(), "sdk-fs-"));
|
||||||
|
const nestedDir = join(directory, "nested");
|
||||||
|
const filePath = join(directory, "notes.txt");
|
||||||
|
const movedPath = join(directory, "notes-moved.txt");
|
||||||
|
const uploadDir = join(directory, "uploaded");
|
||||||
|
|
||||||
|
try {
|
||||||
|
const listedAgents = await sdk.listAgents({ config: true, noCache: true });
|
||||||
|
expect(listedAgents.agents.some((agent) => agent.id === "mock")).toBe(true);
|
||||||
|
|
||||||
|
const mockAgent = await sdk.getAgent("mock", { config: true, noCache: true });
|
||||||
|
expect(mockAgent.id).toBe("mock");
|
||||||
|
expect(Array.isArray(mockAgent.configOptions)).toBe(true);
|
||||||
|
|
||||||
|
await sdk.mkdirFs({ path: nestedDir });
|
||||||
|
await sdk.writeFsFile({ path: filePath }, "hello from sdk");
|
||||||
|
|
||||||
|
const bytes = await sdk.readFsFile({ path: filePath });
|
||||||
|
expect(new TextDecoder().decode(bytes)).toBe("hello from sdk");
|
||||||
|
|
||||||
|
const stat = await sdk.statFs({ path: filePath });
|
||||||
|
expect(stat.path).toBe(filePath);
|
||||||
|
expect(stat.size).toBe(bytes.byteLength);
|
||||||
|
|
||||||
|
const entries = await sdk.listFsEntries({ path: directory });
|
||||||
|
expect(entries.some((entry) => entry.path === nestedDir)).toBe(true);
|
||||||
|
expect(entries.some((entry) => entry.path === filePath)).toBe(true);
|
||||||
|
|
||||||
|
const moved = await sdk.moveFs({
|
||||||
|
from: filePath,
|
||||||
|
to: movedPath,
|
||||||
|
overwrite: true,
|
||||||
|
});
|
||||||
|
expect(moved.to).toBe(movedPath);
|
||||||
|
|
||||||
|
const uploadResult = await sdk.uploadFsBatch(
|
||||||
|
buildTarArchive([{ name: "batch.txt", content: "batch upload works" }]),
|
||||||
|
{ path: uploadDir },
|
||||||
|
);
|
||||||
|
expect(uploadResult.paths.some((path) => path.endsWith("batch.txt"))).toBe(true);
|
||||||
|
|
||||||
|
const uploaded = await sdk.readFsFile({ path: join(uploadDir, "batch.txt") });
|
||||||
|
expect(new TextDecoder().decode(uploaded)).toBe("batch upload works");
|
||||||
|
|
||||||
|
const deleted = await sdk.deleteFsEntry({ path: movedPath });
|
||||||
|
expect(deleted.path).toBe(movedPath);
|
||||||
|
} finally {
|
||||||
|
rmSync(directory, { recursive: true, force: true });
|
||||||
|
await sdk.dispose();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
it("uses custom fetch for both HTTP helpers and ACP session traffic", async () => {
|
it("uses custom fetch for both HTTP helpers and ACP session traffic", async () => {
|
||||||
const defaultFetch = globalThis.fetch;
|
const defaultFetch = globalThis.fetch;
|
||||||
if (!defaultFetch) {
|
if (!defaultFetch) {
|
||||||
|
|
@ -168,7 +329,7 @@ describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
expect(seenPaths.some((path) => path.startsWith("/v1/acp/"))).toBe(true);
|
expect(seenPaths.some((path) => path.startsWith("/v1/acp/"))).toBe(true);
|
||||||
|
|
||||||
await sdk.dispose();
|
await sdk.dispose();
|
||||||
});
|
}, 60_000);
|
||||||
|
|
||||||
it("requires baseUrl when fetch is not provided", async () => {
|
it("requires baseUrl when fetch is not provided", async () => {
|
||||||
await expect(SandboxAgent.connect({ token } as any)).rejects.toThrow(
|
await expect(SandboxAgent.connect({ token } as any)).rejects.toThrow(
|
||||||
|
|
@ -441,4 +602,186 @@ describe("Integration: TypeScript SDK flat session API", () => {
|
||||||
await sdk.dispose();
|
await sdk.dispose();
|
||||||
rmSync(directory, { recursive: true, force: true });
|
rmSync(directory, { recursive: true, force: true });
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it("covers process runtime HTTP helpers, log streaming, and terminal websocket access", async () => {
|
||||||
|
const sdk = await SandboxAgent.connect({
|
||||||
|
baseUrl,
|
||||||
|
token,
|
||||||
|
});
|
||||||
|
|
||||||
|
const originalConfig = await sdk.getProcessConfig();
|
||||||
|
const updatedConfig = await sdk.setProcessConfig({
|
||||||
|
...originalConfig,
|
||||||
|
maxOutputBytes: originalConfig.maxOutputBytes + 1,
|
||||||
|
});
|
||||||
|
expect(updatedConfig.maxOutputBytes).toBe(originalConfig.maxOutputBytes + 1);
|
||||||
|
|
||||||
|
const runResult = await sdk.runProcess({
|
||||||
|
...nodeCommand("process.stdout.write('run-stdout'); process.stderr.write('run-stderr');"),
|
||||||
|
timeoutMs: 5_000,
|
||||||
|
});
|
||||||
|
expect(runResult.stdout).toContain("run-stdout");
|
||||||
|
expect(runResult.stderr).toContain("run-stderr");
|
||||||
|
|
||||||
|
let interactiveProcessId: string | undefined;
|
||||||
|
let ttyProcessId: string | undefined;
|
||||||
|
let killProcessId: string | undefined;
|
||||||
|
|
||||||
|
try {
|
||||||
|
const interactiveProcess = await sdk.createProcess({
|
||||||
|
...nodeCommand(`
|
||||||
|
process.stdin.setEncoding("utf8");
|
||||||
|
process.stdout.write("ready\\n");
|
||||||
|
process.stdin.on("data", (chunk) => {
|
||||||
|
process.stdout.write("echo:" + chunk);
|
||||||
|
});
|
||||||
|
setInterval(() => {}, 1_000);
|
||||||
|
`),
|
||||||
|
interactive: true,
|
||||||
|
});
|
||||||
|
interactiveProcessId = interactiveProcess.id;
|
||||||
|
|
||||||
|
const listed = await sdk.listProcesses();
|
||||||
|
expect(listed.processes.some((process) => process.id === interactiveProcess.id)).toBe(true);
|
||||||
|
|
||||||
|
const fetched = await sdk.getProcess(interactiveProcess.id);
|
||||||
|
expect(fetched.status).toBe("running");
|
||||||
|
|
||||||
|
const initialLogs = await waitForAsync(async () => {
|
||||||
|
const logs = await sdk.getProcessLogs(interactiveProcess.id, { tail: 10 });
|
||||||
|
return logs.entries.some((entry) => decodeProcessLogData(entry.data, entry.encoding).includes("ready"))
|
||||||
|
? logs
|
||||||
|
: undefined;
|
||||||
|
});
|
||||||
|
expect(
|
||||||
|
initialLogs.entries.some((entry) => decodeProcessLogData(entry.data, entry.encoding).includes("ready")),
|
||||||
|
).toBe(true);
|
||||||
|
|
||||||
|
const followedLogs: string[] = [];
|
||||||
|
const subscription = await sdk.followProcessLogs(
|
||||||
|
interactiveProcess.id,
|
||||||
|
(entry) => {
|
||||||
|
followedLogs.push(decodeProcessLogData(entry.data, entry.encoding));
|
||||||
|
},
|
||||||
|
{ tail: 1 },
|
||||||
|
);
|
||||||
|
|
||||||
|
try {
|
||||||
|
const inputResult = await sdk.sendProcessInput(interactiveProcess.id, {
|
||||||
|
data: Buffer.from("hello over stdin\n", "utf8").toString("base64"),
|
||||||
|
encoding: "base64",
|
||||||
|
});
|
||||||
|
expect(inputResult.bytesWritten).toBeGreaterThan(0);
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
const joined = followedLogs.join("");
|
||||||
|
return joined.includes("echo:hello over stdin") ? joined : undefined;
|
||||||
|
});
|
||||||
|
} finally {
|
||||||
|
subscription.close();
|
||||||
|
await subscription.closed;
|
||||||
|
}
|
||||||
|
|
||||||
|
const stopped = await sdk.stopProcess(interactiveProcess.id, { waitMs: 5_000 });
|
||||||
|
expect(stopped.status).toBe("exited");
|
||||||
|
|
||||||
|
await sdk.deleteProcess(interactiveProcess.id);
|
||||||
|
interactiveProcessId = undefined;
|
||||||
|
|
||||||
|
const ttyProcess = await sdk.createProcess({
|
||||||
|
...nodeCommand(`
|
||||||
|
process.stdin.setEncoding("utf8");
|
||||||
|
process.stdin.on("data", (chunk) => {
|
||||||
|
process.stdout.write(chunk);
|
||||||
|
});
|
||||||
|
setInterval(() => {}, 1_000);
|
||||||
|
`),
|
||||||
|
interactive: true,
|
||||||
|
tty: true,
|
||||||
|
});
|
||||||
|
ttyProcessId = ttyProcess.id;
|
||||||
|
|
||||||
|
const resized = await sdk.resizeProcessTerminal(ttyProcess.id, {
|
||||||
|
cols: 120,
|
||||||
|
rows: 40,
|
||||||
|
});
|
||||||
|
expect(resized.cols).toBe(120);
|
||||||
|
expect(resized.rows).toBe(40);
|
||||||
|
|
||||||
|
const wsUrl = sdk.buildProcessTerminalWebSocketUrl(ttyProcess.id);
|
||||||
|
expect(wsUrl.startsWith("ws://") || wsUrl.startsWith("wss://")).toBe(true);
|
||||||
|
|
||||||
|
const ws = sdk.connectProcessTerminalWebSocket(ttyProcess.id, {
|
||||||
|
WebSocket: WebSocket as unknown as typeof globalThis.WebSocket,
|
||||||
|
});
|
||||||
|
ws.binaryType = "arraybuffer";
|
||||||
|
|
||||||
|
const socketTextFrames: string[] = [];
|
||||||
|
const socketBinaryFrames: string[] = [];
|
||||||
|
ws.addEventListener("message", (event) => {
|
||||||
|
if (typeof event.data === "string") {
|
||||||
|
socketTextFrames.push(event.data);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
socketBinaryFrames.push(decodeSocketPayload(event.data));
|
||||||
|
});
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
const ready = socketTextFrames.find((frame) => frame.includes('"type":"ready"'));
|
||||||
|
return ready;
|
||||||
|
});
|
||||||
|
|
||||||
|
ws.send(JSON.stringify({
|
||||||
|
type: "input",
|
||||||
|
data: "hello tty\n",
|
||||||
|
}));
|
||||||
|
|
||||||
|
await waitFor(() => {
|
||||||
|
const joined = socketBinaryFrames.join("");
|
||||||
|
return joined.includes("hello tty") ? joined : undefined;
|
||||||
|
});
|
||||||
|
|
||||||
|
ws.close();
|
||||||
|
await waitForAsync(async () => {
|
||||||
|
const processInfo = await sdk.getProcess(ttyProcess.id);
|
||||||
|
return processInfo.status === "running" ? processInfo : undefined;
|
||||||
|
});
|
||||||
|
|
||||||
|
const killedTty = await sdk.killProcess(ttyProcess.id, { waitMs: 5_000 });
|
||||||
|
expect(killedTty.status).toBe("exited");
|
||||||
|
|
||||||
|
await sdk.deleteProcess(ttyProcess.id);
|
||||||
|
ttyProcessId = undefined;
|
||||||
|
|
||||||
|
const killProcess = await sdk.createProcess({
|
||||||
|
...nodeCommand("setInterval(() => {}, 1_000);"),
|
||||||
|
});
|
||||||
|
killProcessId = killProcess.id;
|
||||||
|
|
||||||
|
const killed = await sdk.killProcess(killProcess.id, { waitMs: 5_000 });
|
||||||
|
expect(killed.status).toBe("exited");
|
||||||
|
|
||||||
|
await sdk.deleteProcess(killProcess.id);
|
||||||
|
killProcessId = undefined;
|
||||||
|
} finally {
|
||||||
|
await sdk.setProcessConfig(originalConfig);
|
||||||
|
|
||||||
|
if (interactiveProcessId) {
|
||||||
|
await sdk.killProcess(interactiveProcessId, { waitMs: 5_000 }).catch(() => {});
|
||||||
|
await sdk.deleteProcess(interactiveProcessId).catch(() => {});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ttyProcessId) {
|
||||||
|
await sdk.killProcess(ttyProcessId, { waitMs: 5_000 }).catch(() => {});
|
||||||
|
await sdk.deleteProcess(ttyProcessId).catch(() => {});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (killProcessId) {
|
||||||
|
await sdk.killProcess(killProcessId, { waitMs: 5_000 }).catch(() => {});
|
||||||
|
await sdk.deleteProcess(killProcessId).catch(() => {});
|
||||||
|
}
|
||||||
|
|
||||||
|
await sdk.dispose();
|
||||||
|
}
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
|
||||||
|
|
@ -1,17 +1,17 @@
|
||||||
# Server Instructions
|
# Server Instructions
|
||||||
|
|
||||||
## ACP v2 Architecture
|
## Architecture
|
||||||
|
|
||||||
- Public API routes are defined in `server/packages/sandbox-agent/src/router.rs`.
|
- Public API routes are defined in `server/packages/sandbox-agent/src/router.rs`.
|
||||||
- ACP runtime/process bridge is in `server/packages/sandbox-agent/src/acp_runtime.rs`.
|
- ACP proxy runtime is in `server/packages/sandbox-agent/src/acp_proxy_runtime.rs`.
|
||||||
- `/v2` is the only active API surface for sessions/prompts (`/v2/rpc`).
|
- All API endpoints are under `/v1`.
|
||||||
- Keep binary filesystem transfer endpoints as dedicated HTTP APIs:
|
- Keep binary filesystem transfer endpoints as dedicated HTTP APIs:
|
||||||
- `GET /v2/fs/file`
|
- `GET /v1/fs/file`
|
||||||
- `PUT /v2/fs/file`
|
- `PUT /v1/fs/file`
|
||||||
- `POST /v2/fs/upload-batch`
|
- `POST /v1/fs/upload-batch`
|
||||||
- Rationale: host-owned cross-agent-consistent behavior and large binary transfer needs that ACP JSON-RPC is not suited to stream efficiently.
|
- Rationale: host-owned cross-agent-consistent behavior and large binary transfer needs that ACP JSON-RPC is not suited to stream efficiently.
|
||||||
- Maintain ACP variants in parallel only when they share the same underlying filesystem implementation; SDK defaults should still prefer HTTP for large/binary transfers.
|
- Maintain ACP variants in parallel only when they share the same underlying filesystem implementation; SDK defaults should still prefer HTTP for large/binary transfers.
|
||||||
- `/v1/*` must remain hard-removed (`410`) and `/opencode/*` stays disabled (`503`) until Phase 7.
|
- `/opencode/*` stays disabled (`503`) until Phase 7.
|
||||||
- Agent install logic (native + ACP agent process + lazy install) is handled by `server/packages/agent-management/`.
|
- Agent install logic (native + ACP agent process + lazy install) is handled by `server/packages/agent-management/`.
|
||||||
|
|
||||||
## API Contract Rules
|
## API Contract Rules
|
||||||
|
|
@ -23,14 +23,14 @@
|
||||||
|
|
||||||
## Tests
|
## Tests
|
||||||
|
|
||||||
Primary v2 integration coverage:
|
Primary v1 integration coverage:
|
||||||
- `server/packages/sandbox-agent/tests/v2_api.rs`
|
- `server/packages/sandbox-agent/tests/v1_api.rs`
|
||||||
- `server/packages/sandbox-agent/tests/v2_agent_process_matrix.rs`
|
- `server/packages/sandbox-agent/tests/v1_agent_process_matrix.rs`
|
||||||
|
|
||||||
Run:
|
Run:
|
||||||
```bash
|
```bash
|
||||||
cargo test -p sandbox-agent --test v2_api
|
cargo test -p sandbox-agent --test v1_api
|
||||||
cargo test -p sandbox-agent --test v2_agent_process_matrix
|
cargo test -p sandbox-agent --test v1_agent_process_matrix
|
||||||
```
|
```
|
||||||
|
|
||||||
## Migration Docs Sync
|
## Migration Docs Sync
|
||||||
|
|
|
||||||
|
|
@ -17,6 +17,7 @@ pub enum ErrorType {
|
||||||
PermissionDenied,
|
PermissionDenied,
|
||||||
NotAcceptable,
|
NotAcceptable,
|
||||||
UnsupportedMediaType,
|
UnsupportedMediaType,
|
||||||
|
NotFound,
|
||||||
SessionNotFound,
|
SessionNotFound,
|
||||||
SessionAlreadyExists,
|
SessionAlreadyExists,
|
||||||
ModeNotSupported,
|
ModeNotSupported,
|
||||||
|
|
@ -37,6 +38,7 @@ impl ErrorType {
|
||||||
Self::PermissionDenied => "urn:sandbox-agent:error:permission_denied",
|
Self::PermissionDenied => "urn:sandbox-agent:error:permission_denied",
|
||||||
Self::NotAcceptable => "urn:sandbox-agent:error:not_acceptable",
|
Self::NotAcceptable => "urn:sandbox-agent:error:not_acceptable",
|
||||||
Self::UnsupportedMediaType => "urn:sandbox-agent:error:unsupported_media_type",
|
Self::UnsupportedMediaType => "urn:sandbox-agent:error:unsupported_media_type",
|
||||||
|
Self::NotFound => "urn:sandbox-agent:error:not_found",
|
||||||
Self::SessionNotFound => "urn:sandbox-agent:error:session_not_found",
|
Self::SessionNotFound => "urn:sandbox-agent:error:session_not_found",
|
||||||
Self::SessionAlreadyExists => "urn:sandbox-agent:error:session_already_exists",
|
Self::SessionAlreadyExists => "urn:sandbox-agent:error:session_already_exists",
|
||||||
Self::ModeNotSupported => "urn:sandbox-agent:error:mode_not_supported",
|
Self::ModeNotSupported => "urn:sandbox-agent:error:mode_not_supported",
|
||||||
|
|
@ -57,6 +59,7 @@ impl ErrorType {
|
||||||
Self::PermissionDenied => "Permission Denied",
|
Self::PermissionDenied => "Permission Denied",
|
||||||
Self::NotAcceptable => "Not Acceptable",
|
Self::NotAcceptable => "Not Acceptable",
|
||||||
Self::UnsupportedMediaType => "Unsupported Media Type",
|
Self::UnsupportedMediaType => "Unsupported Media Type",
|
||||||
|
Self::NotFound => "Not Found",
|
||||||
Self::SessionNotFound => "Session Not Found",
|
Self::SessionNotFound => "Session Not Found",
|
||||||
Self::SessionAlreadyExists => "Session Already Exists",
|
Self::SessionAlreadyExists => "Session Already Exists",
|
||||||
Self::ModeNotSupported => "Mode Not Supported",
|
Self::ModeNotSupported => "Mode Not Supported",
|
||||||
|
|
@ -77,6 +80,7 @@ impl ErrorType {
|
||||||
Self::PermissionDenied => 403,
|
Self::PermissionDenied => 403,
|
||||||
Self::NotAcceptable => 406,
|
Self::NotAcceptable => 406,
|
||||||
Self::UnsupportedMediaType => 415,
|
Self::UnsupportedMediaType => 415,
|
||||||
|
Self::NotFound => 404,
|
||||||
Self::SessionNotFound => 404,
|
Self::SessionNotFound => 404,
|
||||||
Self::SessionAlreadyExists => 409,
|
Self::SessionAlreadyExists => 409,
|
||||||
Self::ModeNotSupported => 400,
|
Self::ModeNotSupported => 400,
|
||||||
|
|
@ -155,6 +159,8 @@ pub enum SandboxError {
|
||||||
NotAcceptable { message: String },
|
NotAcceptable { message: String },
|
||||||
#[error("unsupported media type: {message}")]
|
#[error("unsupported media type: {message}")]
|
||||||
UnsupportedMediaType { message: String },
|
UnsupportedMediaType { message: String },
|
||||||
|
#[error("not found: {resource} {id}")]
|
||||||
|
NotFound { resource: String, id: String },
|
||||||
#[error("session not found: {session_id}")]
|
#[error("session not found: {session_id}")]
|
||||||
SessionNotFound { session_id: String },
|
SessionNotFound { session_id: String },
|
||||||
#[error("session already exists: {session_id}")]
|
#[error("session already exists: {session_id}")]
|
||||||
|
|
@ -180,6 +186,7 @@ impl SandboxError {
|
||||||
Self::PermissionDenied { .. } => ErrorType::PermissionDenied,
|
Self::PermissionDenied { .. } => ErrorType::PermissionDenied,
|
||||||
Self::NotAcceptable { .. } => ErrorType::NotAcceptable,
|
Self::NotAcceptable { .. } => ErrorType::NotAcceptable,
|
||||||
Self::UnsupportedMediaType { .. } => ErrorType::UnsupportedMediaType,
|
Self::UnsupportedMediaType { .. } => ErrorType::UnsupportedMediaType,
|
||||||
|
Self::NotFound { .. } => ErrorType::NotFound,
|
||||||
Self::SessionNotFound { .. } => ErrorType::SessionNotFound,
|
Self::SessionNotFound { .. } => ErrorType::SessionNotFound,
|
||||||
Self::SessionAlreadyExists { .. } => ErrorType::SessionAlreadyExists,
|
Self::SessionAlreadyExists { .. } => ErrorType::SessionAlreadyExists,
|
||||||
Self::ModeNotSupported { .. } => ErrorType::ModeNotSupported,
|
Self::ModeNotSupported { .. } => ErrorType::ModeNotSupported,
|
||||||
|
|
@ -264,6 +271,12 @@ impl SandboxError {
|
||||||
map.insert("message".to_string(), Value::String(message.clone()));
|
map.insert("message".to_string(), Value::String(message.clone()));
|
||||||
(None, None, Some(Value::Object(map)))
|
(None, None, Some(Value::Object(map)))
|
||||||
}
|
}
|
||||||
|
Self::NotFound { resource, id } => {
|
||||||
|
let mut map = Map::new();
|
||||||
|
map.insert("resource".to_string(), Value::String(resource.clone()));
|
||||||
|
map.insert("id".to_string(), Value::String(id.clone()));
|
||||||
|
(None, None, Some(Value::Object(map)))
|
||||||
|
}
|
||||||
Self::SessionNotFound { session_id } => (None, Some(session_id.clone()), None),
|
Self::SessionNotFound { session_id } => (None, Some(session_id.clone()), None),
|
||||||
Self::SessionAlreadyExists { session_id } => (None, Some(session_id.clone()), None),
|
Self::SessionAlreadyExists { session_id } => (None, Some(session_id.clone()), None),
|
||||||
Self::ModeNotSupported { agent, mode } => {
|
Self::ModeNotSupported { agent, mode } => {
|
||||||
|
|
|
||||||
|
|
@ -55,6 +55,7 @@ insta.workspace = true
|
||||||
tower.workspace = true
|
tower.workspace = true
|
||||||
tempfile.workspace = true
|
tempfile.workspace = true
|
||||||
serial_test = "3.2"
|
serial_test = "3.2"
|
||||||
|
tokio-tungstenite = "0.24"
|
||||||
|
|
||||||
[features]
|
[features]
|
||||||
test-utils = ["tempfile"]
|
test-utils = ["tempfile"]
|
||||||
|
|
|
||||||
|
|
@ -3,6 +3,7 @@
|
||||||
mod acp_proxy_runtime;
|
mod acp_proxy_runtime;
|
||||||
pub mod cli;
|
pub mod cli;
|
||||||
pub mod daemon;
|
pub mod daemon;
|
||||||
|
mod process_runtime;
|
||||||
pub mod router;
|
pub mod router;
|
||||||
pub mod server_logs;
|
pub mod server_logs;
|
||||||
pub mod telemetry;
|
pub mod telemetry;
|
||||||
|
|
|
||||||
1082
server/packages/sandbox-agent/src/process_runtime.rs
Normal file
1082
server/packages/sandbox-agent/src/process_runtime.rs
Normal file
File diff suppressed because it is too large
Load diff
|
|
@ -1,4 +1,5 @@
|
||||||
use std::collections::{BTreeMap, HashMap};
|
use std::collections::{BTreeMap, HashMap};
|
||||||
|
use std::convert::Infallible;
|
||||||
use std::fs;
|
use std::fs;
|
||||||
use std::io::Cursor;
|
use std::io::Cursor;
|
||||||
use std::path::{Path as StdPath, PathBuf};
|
use std::path::{Path as StdPath, PathBuf};
|
||||||
|
|
@ -6,6 +7,7 @@ use std::sync::{Arc, Mutex};
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
|
|
||||||
use axum::body::Bytes;
|
use axum::body::Bytes;
|
||||||
|
use axum::extract::ws::{Message, WebSocket, WebSocketUpgrade};
|
||||||
use axum::extract::{Path, Query, State};
|
use axum::extract::{Path, Query, State};
|
||||||
use axum::http::{header, HeaderMap, Request, StatusCode};
|
use axum::http::{header, HeaderMap, Request, StatusCode};
|
||||||
use axum::middleware::Next;
|
use axum::middleware::Next;
|
||||||
|
|
@ -13,6 +15,8 @@ use axum::response::sse::KeepAlive;
|
||||||
use axum::response::{IntoResponse, Response, Sse};
|
use axum::response::{IntoResponse, Response, Sse};
|
||||||
use axum::routing::{delete, get, post};
|
use axum::routing::{delete, get, post};
|
||||||
use axum::{Json, Router};
|
use axum::{Json, Router};
|
||||||
|
use futures::stream;
|
||||||
|
use futures::StreamExt;
|
||||||
use sandbox_agent_agent_management::agents::{
|
use sandbox_agent_agent_management::agents::{
|
||||||
AgentId, AgentManager, InstallOptions, InstallResult, InstallSource, InstalledArtifactKind,
|
AgentId, AgentManager, InstallOptions, InstallResult, InstallSource, InstalledArtifactKind,
|
||||||
};
|
};
|
||||||
|
|
@ -27,11 +31,16 @@ use serde::de::DeserializeOwned;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use serde_json::{json, Value};
|
use serde_json::{json, Value};
|
||||||
use tar::Archive;
|
use tar::Archive;
|
||||||
|
use tokio_stream::wrappers::BroadcastStream;
|
||||||
use tower_http::trace::TraceLayer;
|
use tower_http::trace::TraceLayer;
|
||||||
use tracing::Span;
|
use tracing::Span;
|
||||||
use utoipa::{Modify, OpenApi, ToSchema};
|
use utoipa::{Modify, OpenApi, ToSchema};
|
||||||
|
|
||||||
use crate::acp_proxy_runtime::{AcpProxyRuntime, ProxyPostOutcome};
|
use crate::acp_proxy_runtime::{AcpProxyRuntime, ProxyPostOutcome};
|
||||||
|
use crate::process_runtime::{
|
||||||
|
decode_input_bytes, ProcessLogFilter, ProcessLogFilterStream, ProcessRuntime,
|
||||||
|
ProcessRuntimeConfig, ProcessSnapshot, ProcessStartSpec, ProcessStatus, ProcessStream, RunSpec,
|
||||||
|
};
|
||||||
use crate::ui;
|
use crate::ui;
|
||||||
|
|
||||||
mod support;
|
mod support;
|
||||||
|
|
@ -77,6 +86,7 @@ pub struct AppState {
|
||||||
agent_manager: Arc<AgentManager>,
|
agent_manager: Arc<AgentManager>,
|
||||||
acp_proxy: Arc<AcpProxyRuntime>,
|
acp_proxy: Arc<AcpProxyRuntime>,
|
||||||
opencode_server_manager: Arc<OpenCodeServerManager>,
|
opencode_server_manager: Arc<OpenCodeServerManager>,
|
||||||
|
process_runtime: Arc<ProcessRuntime>,
|
||||||
pub(crate) branding: BrandingMode,
|
pub(crate) branding: BrandingMode,
|
||||||
version_cache: Mutex<HashMap<AgentId, CachedAgentVersion>>,
|
version_cache: Mutex<HashMap<AgentId, CachedAgentVersion>>,
|
||||||
}
|
}
|
||||||
|
|
@ -100,11 +110,13 @@ impl AppState {
|
||||||
auto_restart: true,
|
auto_restart: true,
|
||||||
},
|
},
|
||||||
));
|
));
|
||||||
|
let process_runtime = Arc::new(ProcessRuntime::new());
|
||||||
Self {
|
Self {
|
||||||
auth,
|
auth,
|
||||||
agent_manager,
|
agent_manager,
|
||||||
acp_proxy,
|
acp_proxy,
|
||||||
opencode_server_manager,
|
opencode_server_manager,
|
||||||
|
process_runtime,
|
||||||
branding,
|
branding,
|
||||||
version_cache: Mutex::new(HashMap::new()),
|
version_cache: Mutex::new(HashMap::new()),
|
||||||
}
|
}
|
||||||
|
|
@ -122,6 +134,10 @@ impl AppState {
|
||||||
self.opencode_server_manager.clone()
|
self.opencode_server_manager.clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub(crate) fn process_runtime(&self) -> Arc<ProcessRuntime> {
|
||||||
|
self.process_runtime.clone()
|
||||||
|
}
|
||||||
|
|
||||||
pub(crate) fn purge_version_cache(&self, agent: AgentId) {
|
pub(crate) fn purge_version_cache(&self, agent: AgentId) {
|
||||||
self.version_cache.lock().unwrap().remove(&agent);
|
self.version_cache.lock().unwrap().remove(&agent);
|
||||||
}
|
}
|
||||||
|
|
@ -166,6 +182,28 @@ pub fn build_router_with_state(shared: Arc<AppState>) -> (Router, Arc<AppState>)
|
||||||
.route("/fs/move", post(post_v1_fs_move))
|
.route("/fs/move", post(post_v1_fs_move))
|
||||||
.route("/fs/stat", get(get_v1_fs_stat))
|
.route("/fs/stat", get(get_v1_fs_stat))
|
||||||
.route("/fs/upload-batch", post(post_v1_fs_upload_batch))
|
.route("/fs/upload-batch", post(post_v1_fs_upload_batch))
|
||||||
|
.route(
|
||||||
|
"/processes/config",
|
||||||
|
get(get_v1_processes_config).post(post_v1_processes_config),
|
||||||
|
)
|
||||||
|
.route("/processes", get(get_v1_processes).post(post_v1_processes))
|
||||||
|
.route("/processes/run", post(post_v1_processes_run))
|
||||||
|
.route(
|
||||||
|
"/processes/:id",
|
||||||
|
get(get_v1_process).delete(delete_v1_process),
|
||||||
|
)
|
||||||
|
.route("/processes/:id/stop", post(post_v1_process_stop))
|
||||||
|
.route("/processes/:id/kill", post(post_v1_process_kill))
|
||||||
|
.route("/processes/:id/logs", get(get_v1_process_logs))
|
||||||
|
.route("/processes/:id/input", post(post_v1_process_input))
|
||||||
|
.route(
|
||||||
|
"/processes/:id/terminal/resize",
|
||||||
|
post(post_v1_process_terminal_resize),
|
||||||
|
)
|
||||||
|
.route(
|
||||||
|
"/processes/:id/terminal/ws",
|
||||||
|
get(get_v1_process_terminal_ws),
|
||||||
|
)
|
||||||
.route(
|
.route(
|
||||||
"/config/mcp",
|
"/config/mcp",
|
||||||
get(get_v1_config_mcp)
|
get(get_v1_config_mcp)
|
||||||
|
|
@ -295,6 +333,19 @@ pub async fn shutdown_servers(state: &Arc<AppState>) {
|
||||||
post_v1_fs_move,
|
post_v1_fs_move,
|
||||||
get_v1_fs_stat,
|
get_v1_fs_stat,
|
||||||
post_v1_fs_upload_batch,
|
post_v1_fs_upload_batch,
|
||||||
|
get_v1_processes_config,
|
||||||
|
post_v1_processes_config,
|
||||||
|
post_v1_processes,
|
||||||
|
post_v1_processes_run,
|
||||||
|
get_v1_processes,
|
||||||
|
get_v1_process,
|
||||||
|
post_v1_process_stop,
|
||||||
|
post_v1_process_kill,
|
||||||
|
delete_v1_process,
|
||||||
|
get_v1_process_logs,
|
||||||
|
post_v1_process_input,
|
||||||
|
post_v1_process_terminal_resize,
|
||||||
|
get_v1_process_terminal_ws,
|
||||||
get_v1_config_mcp,
|
get_v1_config_mcp,
|
||||||
put_v1_config_mcp,
|
put_v1_config_mcp,
|
||||||
delete_v1_config_mcp,
|
delete_v1_config_mcp,
|
||||||
|
|
@ -329,6 +380,22 @@ pub async fn shutdown_servers(state: &Arc<AppState>) {
|
||||||
FsMoveResponse,
|
FsMoveResponse,
|
||||||
FsActionResponse,
|
FsActionResponse,
|
||||||
FsUploadBatchResponse,
|
FsUploadBatchResponse,
|
||||||
|
ProcessConfig,
|
||||||
|
ProcessCreateRequest,
|
||||||
|
ProcessRunRequest,
|
||||||
|
ProcessRunResponse,
|
||||||
|
ProcessState,
|
||||||
|
ProcessInfo,
|
||||||
|
ProcessListResponse,
|
||||||
|
ProcessLogsStream,
|
||||||
|
ProcessLogsQuery,
|
||||||
|
ProcessLogEntry,
|
||||||
|
ProcessLogsResponse,
|
||||||
|
ProcessInputRequest,
|
||||||
|
ProcessInputResponse,
|
||||||
|
ProcessSignalQuery,
|
||||||
|
ProcessTerminalResizeRequest,
|
||||||
|
ProcessTerminalResizeResponse,
|
||||||
AcpPostQuery,
|
AcpPostQuery,
|
||||||
AcpServerInfo,
|
AcpServerInfo,
|
||||||
AcpServerListResponse,
|
AcpServerListResponse,
|
||||||
|
|
@ -361,12 +428,21 @@ impl Modify for ServerAddon {
|
||||||
pub enum ApiError {
|
pub enum ApiError {
|
||||||
#[error(transparent)]
|
#[error(transparent)]
|
||||||
Sandbox(#[from] SandboxError),
|
Sandbox(#[from] SandboxError),
|
||||||
|
#[error("problem: {0:?}")]
|
||||||
|
Problem(ProblemDetails),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl From<ProblemDetails> for ApiError {
|
||||||
|
fn from(value: ProblemDetails) -> Self {
|
||||||
|
Self::Problem(value)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl IntoResponse for ApiError {
|
impl IntoResponse for ApiError {
|
||||||
fn into_response(self) -> Response {
|
fn into_response(self) -> Response {
|
||||||
let problem = match &self {
|
let problem = match &self {
|
||||||
ApiError::Sandbox(error) => problem_from_sandbox_error(error),
|
ApiError::Sandbox(error) => problem_from_sandbox_error(error),
|
||||||
|
ApiError::Problem(problem) => problem.clone(),
|
||||||
};
|
};
|
||||||
let status =
|
let status =
|
||||||
StatusCode::from_u16(problem.status).unwrap_or(StatusCode::INTERNAL_SERVER_ERROR);
|
StatusCode::from_u16(problem.status).unwrap_or(StatusCode::INTERNAL_SERVER_ERROR);
|
||||||
|
|
@ -1075,6 +1151,678 @@ async fn post_v1_fs_upload_batch(
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Get process runtime configuration.
|
||||||
|
///
|
||||||
|
/// Returns the current runtime configuration for the process management API,
|
||||||
|
/// including limits for concurrency, timeouts, and buffer sizes.
|
||||||
|
#[utoipa::path(
|
||||||
|
get,
|
||||||
|
path = "/v1/processes/config",
|
||||||
|
tag = "v1",
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Current runtime process config", body = ProcessConfig),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn get_v1_processes_config(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
) -> Result<Json<ProcessConfig>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let config = state.process_runtime().get_config().await;
|
||||||
|
Ok(Json(map_process_config(config)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Update process runtime configuration.
|
||||||
|
///
|
||||||
|
/// Replaces the runtime configuration for the process management API.
|
||||||
|
/// Validates that all values are non-zero and clamps default timeout to max.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/config",
|
||||||
|
tag = "v1",
|
||||||
|
request_body = ProcessConfig,
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Updated runtime process config", body = ProcessConfig),
|
||||||
|
(status = 400, description = "Invalid config", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_processes_config(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Json(body): Json<ProcessConfig>,
|
||||||
|
) -> Result<Json<ProcessConfig>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
let updated = runtime
|
||||||
|
.set_config(into_runtime_process_config(body))
|
||||||
|
.await?;
|
||||||
|
Ok(Json(map_process_config(updated)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a long-lived managed process.
|
||||||
|
///
|
||||||
|
/// Spawns a new process with the given command and arguments. Supports both
|
||||||
|
/// pipe-based and PTY (tty) modes. Returns the process descriptor on success.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes",
|
||||||
|
tag = "v1",
|
||||||
|
request_body = ProcessCreateRequest,
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Started process", body = ProcessInfo),
|
||||||
|
(status = 400, description = "Invalid request", body = ProblemDetails),
|
||||||
|
(status = 409, description = "Process limit or state conflict", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_processes(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Json(body): Json<ProcessCreateRequest>,
|
||||||
|
) -> Result<Json<ProcessInfo>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
let snapshot = runtime
|
||||||
|
.start_process(ProcessStartSpec {
|
||||||
|
command: body.command,
|
||||||
|
args: body.args,
|
||||||
|
cwd: body.cwd,
|
||||||
|
env: body.env.into_iter().collect(),
|
||||||
|
tty: body.tty,
|
||||||
|
interactive: body.interactive,
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(Json(map_process_snapshot(snapshot)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Run a one-shot command.
|
||||||
|
///
|
||||||
|
/// Executes a command to completion and returns its stdout, stderr, exit code,
|
||||||
|
/// and duration. Supports configurable timeout and output size limits.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/run",
|
||||||
|
tag = "v1",
|
||||||
|
request_body = ProcessRunRequest,
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "One-off command result", body = ProcessRunResponse),
|
||||||
|
(status = 400, description = "Invalid request", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_processes_run(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Json(body): Json<ProcessRunRequest>,
|
||||||
|
) -> Result<Json<ProcessRunResponse>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
let output = runtime
|
||||||
|
.run_once(RunSpec {
|
||||||
|
command: body.command,
|
||||||
|
args: body.args,
|
||||||
|
cwd: body.cwd,
|
||||||
|
env: body.env.into_iter().collect(),
|
||||||
|
timeout_ms: body.timeout_ms,
|
||||||
|
max_output_bytes: body.max_output_bytes,
|
||||||
|
})
|
||||||
|
.await?;
|
||||||
|
|
||||||
|
Ok(Json(ProcessRunResponse {
|
||||||
|
exit_code: output.exit_code,
|
||||||
|
timed_out: output.timed_out,
|
||||||
|
stdout: output.stdout,
|
||||||
|
stderr: output.stderr,
|
||||||
|
stdout_truncated: output.stdout_truncated,
|
||||||
|
stderr_truncated: output.stderr_truncated,
|
||||||
|
duration_ms: output.duration_ms,
|
||||||
|
}))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// List all managed processes.
|
||||||
|
///
|
||||||
|
/// Returns a list of all processes (running and exited) currently tracked
|
||||||
|
/// by the runtime, sorted by process ID.
|
||||||
|
#[utoipa::path(
|
||||||
|
get,
|
||||||
|
path = "/v1/processes",
|
||||||
|
tag = "v1",
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "List processes", body = ProcessListResponse),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn get_v1_processes(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
) -> Result<Json<ProcessListResponse>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let snapshots = state.process_runtime().list_processes().await;
|
||||||
|
Ok(Json(ProcessListResponse {
|
||||||
|
processes: snapshots.into_iter().map(map_process_snapshot).collect(),
|
||||||
|
}))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a single process by ID.
|
||||||
|
///
|
||||||
|
/// Returns the current state of a managed process including its status,
|
||||||
|
/// PID, exit code, and creation/exit timestamps.
|
||||||
|
#[utoipa::path(
|
||||||
|
get,
|
||||||
|
path = "/v1/processes/{id}",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Process details", body = ProcessInfo),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn get_v1_process(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
) -> Result<Json<ProcessInfo>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let snapshot = state.process_runtime().snapshot(&id).await?;
|
||||||
|
Ok(Json(map_process_snapshot(snapshot)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Send SIGTERM to a process.
|
||||||
|
///
|
||||||
|
/// Sends SIGTERM to the process and optionally waits up to `waitMs`
|
||||||
|
/// milliseconds for the process to exit before returning.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/{id}/stop",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID"),
|
||||||
|
("waitMs" = Option<u64>, Query, description = "Wait up to N ms for process to exit")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Stop signal sent", body = ProcessInfo),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_process_stop(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
Query(query): Query<ProcessSignalQuery>,
|
||||||
|
) -> Result<Json<ProcessInfo>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let snapshot = state
|
||||||
|
.process_runtime()
|
||||||
|
.stop_process(&id, query.wait_ms)
|
||||||
|
.await?;
|
||||||
|
Ok(Json(map_process_snapshot(snapshot)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Send SIGKILL to a process.
|
||||||
|
///
|
||||||
|
/// Sends SIGKILL to the process and optionally waits up to `waitMs`
|
||||||
|
/// milliseconds for the process to exit before returning.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/{id}/kill",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID"),
|
||||||
|
("waitMs" = Option<u64>, Query, description = "Wait up to N ms for process to exit")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Kill signal sent", body = ProcessInfo),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_process_kill(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
Query(query): Query<ProcessSignalQuery>,
|
||||||
|
) -> Result<Json<ProcessInfo>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let snapshot = state
|
||||||
|
.process_runtime()
|
||||||
|
.kill_process(&id, query.wait_ms)
|
||||||
|
.await?;
|
||||||
|
Ok(Json(map_process_snapshot(snapshot)))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Delete a process record.
|
||||||
|
///
|
||||||
|
/// Removes a stopped process from the runtime. Returns 409 if the process
|
||||||
|
/// is still running; stop or kill it first.
|
||||||
|
#[utoipa::path(
|
||||||
|
delete,
|
||||||
|
path = "/v1/processes/{id}",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 204, description = "Process deleted"),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 409, description = "Process is still running", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn delete_v1_process(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
) -> Result<StatusCode, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
state.process_runtime().delete_process(&id).await?;
|
||||||
|
Ok(StatusCode::NO_CONTENT)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Fetch process logs.
|
||||||
|
///
|
||||||
|
/// Returns buffered log entries for a process. Supports filtering by stream
|
||||||
|
/// type, tail count, and sequence-based resumption. When `follow=true`,
|
||||||
|
/// returns an SSE stream that replays buffered entries then streams live output.
|
||||||
|
#[utoipa::path(
|
||||||
|
get,
|
||||||
|
path = "/v1/processes/{id}/logs",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID"),
|
||||||
|
("stream" = Option<ProcessLogsStream>, Query, description = "stdout|stderr|combined|pty"),
|
||||||
|
("tail" = Option<usize>, Query, description = "Tail N entries"),
|
||||||
|
("follow" = Option<bool>, Query, description = "Follow via SSE"),
|
||||||
|
("since" = Option<u64>, Query, description = "Only entries with sequence greater than this")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Process logs", body = ProcessLogsResponse),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn get_v1_process_logs(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
headers: HeaderMap,
|
||||||
|
Query(query): Query<ProcessLogsQuery>,
|
||||||
|
) -> Result<Response, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
let default_stream = if runtime.is_tty(&id).await? {
|
||||||
|
ProcessLogsStream::Pty
|
||||||
|
} else {
|
||||||
|
ProcessLogsStream::Combined
|
||||||
|
};
|
||||||
|
let requested_stream = query.stream.unwrap_or(default_stream);
|
||||||
|
let since = match (query.since, parse_last_event_id(&headers)?) {
|
||||||
|
(Some(query_since), Some(last_event_id)) => Some(query_since.max(last_event_id)),
|
||||||
|
(Some(query_since), None) => Some(query_since),
|
||||||
|
(None, Some(last_event_id)) => Some(last_event_id),
|
||||||
|
(None, None) => None,
|
||||||
|
};
|
||||||
|
let filter = ProcessLogFilter {
|
||||||
|
stream: into_runtime_log_stream(requested_stream),
|
||||||
|
tail: query.tail,
|
||||||
|
since,
|
||||||
|
};
|
||||||
|
|
||||||
|
let entries = runtime.logs(&id, filter).await?;
|
||||||
|
let response_entries: Vec<ProcessLogEntry> =
|
||||||
|
entries.iter().cloned().map(map_process_log_line).collect();
|
||||||
|
|
||||||
|
if query.follow.unwrap_or(false) {
|
||||||
|
let rx = runtime.subscribe_logs(&id).await?;
|
||||||
|
let replay_stream = stream::iter(response_entries.into_iter().map(|entry| {
|
||||||
|
Ok::<axum::response::sse::Event, Infallible>(
|
||||||
|
axum::response::sse::Event::default()
|
||||||
|
.event("log")
|
||||||
|
.id(entry.sequence.to_string())
|
||||||
|
.data(serde_json::to_string(&entry).unwrap_or_else(|_| "{}".to_string())),
|
||||||
|
)
|
||||||
|
}));
|
||||||
|
|
||||||
|
let requested_stream_copy = requested_stream;
|
||||||
|
let follow_stream = BroadcastStream::new(rx).filter_map(move |item| {
|
||||||
|
let requested_stream_copy = requested_stream_copy;
|
||||||
|
async move {
|
||||||
|
match item {
|
||||||
|
Ok(line) => {
|
||||||
|
let entry = map_process_log_line(line);
|
||||||
|
if process_log_matches(&entry, requested_stream_copy) {
|
||||||
|
Some(Ok(axum::response::sse::Event::default()
|
||||||
|
.event("log")
|
||||||
|
.id(entry.sequence.to_string())
|
||||||
|
.data(
|
||||||
|
serde_json::to_string(&entry)
|
||||||
|
.unwrap_or_else(|_| "{}".to_string()),
|
||||||
|
)))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(_) => None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
let stream = replay_stream.chain(follow_stream);
|
||||||
|
let response =
|
||||||
|
Sse::new(stream).keep_alive(KeepAlive::new().interval(Duration::from_secs(15)));
|
||||||
|
return Ok(response.into_response());
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(Json(ProcessLogsResponse {
|
||||||
|
process_id: id,
|
||||||
|
stream: requested_stream,
|
||||||
|
entries: response_entries,
|
||||||
|
})
|
||||||
|
.into_response())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Write input to a process.
|
||||||
|
///
|
||||||
|
/// Sends data to a process's stdin (pipe mode) or PTY writer (tty mode).
|
||||||
|
/// Data can be encoded as base64, utf8, or text. Returns 413 if the decoded
|
||||||
|
/// payload exceeds the configured `maxInputBytesPerRequest` limit.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/{id}/input",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID")
|
||||||
|
),
|
||||||
|
request_body = ProcessInputRequest,
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Input accepted", body = ProcessInputResponse),
|
||||||
|
(status = 400, description = "Invalid request", body = ProblemDetails),
|
||||||
|
(status = 413, description = "Input exceeds configured limit", body = ProblemDetails),
|
||||||
|
(status = 409, description = "Process not writable", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_process_input(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
Json(body): Json<ProcessInputRequest>,
|
||||||
|
) -> Result<Json<ProcessInputResponse>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let encoding = body.encoding.unwrap_or_else(|| "base64".to_string());
|
||||||
|
let input = decode_input_bytes(&body.data, &encoding)?;
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
let max_input = runtime.max_input_bytes().await;
|
||||||
|
if input.len() > max_input {
|
||||||
|
return Err(SandboxError::InvalidRequest {
|
||||||
|
message: format!("input payload exceeds maxInputBytesPerRequest ({max_input})"),
|
||||||
|
}
|
||||||
|
.into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let bytes_written = runtime.write_input(&id, &input).await?;
|
||||||
|
Ok(Json(ProcessInputResponse { bytes_written }))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Resize a process terminal.
|
||||||
|
///
|
||||||
|
/// Sets the PTY window size (columns and rows) for a tty-mode process and
|
||||||
|
/// sends SIGWINCH so the child process can adapt.
|
||||||
|
#[utoipa::path(
|
||||||
|
post,
|
||||||
|
path = "/v1/processes/{id}/terminal/resize",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID")
|
||||||
|
),
|
||||||
|
request_body = ProcessTerminalResizeRequest,
|
||||||
|
responses(
|
||||||
|
(status = 200, description = "Resize accepted", body = ProcessTerminalResizeResponse),
|
||||||
|
(status = 400, description = "Invalid request", body = ProblemDetails),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 409, description = "Not a terminal process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn post_v1_process_terminal_resize(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
Json(body): Json<ProcessTerminalResizeRequest>,
|
||||||
|
) -> Result<Json<ProcessTerminalResizeResponse>, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
state
|
||||||
|
.process_runtime()
|
||||||
|
.resize_terminal(&id, body.cols, body.rows)
|
||||||
|
.await?;
|
||||||
|
Ok(Json(ProcessTerminalResizeResponse {
|
||||||
|
cols: body.cols,
|
||||||
|
rows: body.rows,
|
||||||
|
}))
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Open an interactive WebSocket terminal session.
|
||||||
|
///
|
||||||
|
/// Upgrades the connection to a WebSocket for bidirectional PTY I/O. Accepts
|
||||||
|
/// `access_token` query param for browser-based auth (WebSocket API cannot
|
||||||
|
/// send custom headers). Streams raw PTY output as binary frames and accepts
|
||||||
|
/// JSON control frames for input, resize, and close.
|
||||||
|
#[utoipa::path(
|
||||||
|
get,
|
||||||
|
path = "/v1/processes/{id}/terminal/ws",
|
||||||
|
tag = "v1",
|
||||||
|
params(
|
||||||
|
("id" = String, Path, description = "Process ID"),
|
||||||
|
("access_token" = Option<String>, Query, description = "Bearer token alternative for WS auth")
|
||||||
|
),
|
||||||
|
responses(
|
||||||
|
(status = 101, description = "WebSocket upgraded"),
|
||||||
|
(status = 400, description = "Invalid websocket frame or upgrade request", body = ProblemDetails),
|
||||||
|
(status = 404, description = "Unknown process", body = ProblemDetails),
|
||||||
|
(status = 409, description = "Not a terminal process", body = ProblemDetails),
|
||||||
|
(status = 501, description = "Process API unsupported on this platform", body = ProblemDetails)
|
||||||
|
)
|
||||||
|
)]
|
||||||
|
async fn get_v1_process_terminal_ws(
|
||||||
|
State(state): State<Arc<AppState>>,
|
||||||
|
Path(id): Path<String>,
|
||||||
|
Query(_query): Query<ProcessWsQuery>,
|
||||||
|
ws: WebSocketUpgrade,
|
||||||
|
) -> Result<Response, ApiError> {
|
||||||
|
if !process_api_supported() {
|
||||||
|
return Err(process_api_not_supported().into());
|
||||||
|
}
|
||||||
|
|
||||||
|
let runtime = state.process_runtime();
|
||||||
|
if !runtime.is_tty(&id).await? {
|
||||||
|
return Err(SandboxError::Conflict {
|
||||||
|
message: "process is not running in tty mode".to_string(),
|
||||||
|
}
|
||||||
|
.into());
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(ws
|
||||||
|
.on_upgrade(move |socket| process_terminal_ws_session(socket, runtime, id))
|
||||||
|
.into_response())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Deserialize)]
|
||||||
|
#[serde(tag = "type", rename_all = "camelCase")]
|
||||||
|
enum TerminalClientFrame {
|
||||||
|
Input {
|
||||||
|
data: String,
|
||||||
|
#[serde(default)]
|
||||||
|
encoding: Option<String>,
|
||||||
|
},
|
||||||
|
Resize {
|
||||||
|
cols: u16,
|
||||||
|
rows: u16,
|
||||||
|
},
|
||||||
|
Close,
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn process_terminal_ws_session(
|
||||||
|
mut socket: WebSocket,
|
||||||
|
runtime: Arc<ProcessRuntime>,
|
||||||
|
id: String,
|
||||||
|
) {
|
||||||
|
let _ = send_ws_json(
|
||||||
|
&mut socket,
|
||||||
|
json!({
|
||||||
|
"type": "ready",
|
||||||
|
"processId": &id,
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
|
||||||
|
let mut log_rx = match runtime.subscribe_logs(&id).await {
|
||||||
|
Ok(rx) => rx,
|
||||||
|
Err(err) => {
|
||||||
|
let _ = send_ws_error(&mut socket, &err.to_string()).await;
|
||||||
|
let _ = socket.close().await;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let mut exit_poll = tokio::time::interval(Duration::from_millis(150));
|
||||||
|
|
||||||
|
loop {
|
||||||
|
tokio::select! {
|
||||||
|
ws_in = socket.recv() => {
|
||||||
|
match ws_in {
|
||||||
|
Some(Ok(Message::Binary(_))) => {
|
||||||
|
let _ = send_ws_error(&mut socket, "binary input is not supported; use text JSON frames").await;
|
||||||
|
}
|
||||||
|
Some(Ok(Message::Text(text))) => {
|
||||||
|
let parsed = serde_json::from_str::<TerminalClientFrame>(&text);
|
||||||
|
match parsed {
|
||||||
|
Ok(TerminalClientFrame::Input { data, encoding }) => {
|
||||||
|
let input = match decode_input_bytes(&data, encoding.as_deref().unwrap_or("utf8")) {
|
||||||
|
Ok(input) => input,
|
||||||
|
Err(err) => {
|
||||||
|
let _ = send_ws_error(&mut socket, &err.to_string()).await;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
let max_input = runtime.max_input_bytes().await;
|
||||||
|
if input.len() > max_input {
|
||||||
|
let _ = send_ws_error(&mut socket, &format!("input payload exceeds maxInputBytesPerRequest ({max_input})")).await;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
if let Err(err) = runtime.write_input(&id, &input).await {
|
||||||
|
let _ = send_ws_error(&mut socket, &err.to_string()).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(TerminalClientFrame::Resize { cols, rows }) => {
|
||||||
|
if let Err(err) = runtime.resize_terminal(&id, cols, rows).await {
|
||||||
|
let _ = send_ws_error(&mut socket, &err.to_string()).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(TerminalClientFrame::Close) => {
|
||||||
|
let _ = socket.close().await;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
Err(err) => {
|
||||||
|
let _ = send_ws_error(&mut socket, &format!("invalid terminal frame: {err}")).await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Some(Ok(Message::Ping(payload))) => {
|
||||||
|
let _ = socket.send(Message::Pong(payload)).await;
|
||||||
|
}
|
||||||
|
Some(Ok(Message::Close(_))) | None => break,
|
||||||
|
Some(Ok(Message::Pong(_))) => {}
|
||||||
|
Some(Err(_)) => break,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
log_in = log_rx.recv() => {
|
||||||
|
match log_in {
|
||||||
|
Ok(line) => {
|
||||||
|
if line.stream != ProcessStream::Pty {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
let bytes = {
|
||||||
|
use base64::engine::general_purpose::STANDARD as BASE64_ENGINE;
|
||||||
|
use base64::Engine;
|
||||||
|
BASE64_ENGINE.decode(&line.data).unwrap_or_default()
|
||||||
|
};
|
||||||
|
if socket.send(Message::Binary(bytes)).await.is_err() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Err(tokio::sync::broadcast::error::RecvError::Lagged(_)) => {}
|
||||||
|
Err(tokio::sync::broadcast::error::RecvError::Closed) => break,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
_ = exit_poll.tick() => {
|
||||||
|
if let Ok(snapshot) = runtime.snapshot(&id).await {
|
||||||
|
if snapshot.status == ProcessStatus::Exited {
|
||||||
|
let _ = send_ws_json(
|
||||||
|
&mut socket,
|
||||||
|
json!({
|
||||||
|
"type": "exit",
|
||||||
|
"exitCode": snapshot.exit_code,
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
let _ = socket.close().await;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn send_ws_json(socket: &mut WebSocket, payload: Value) -> Result<(), ()> {
|
||||||
|
socket
|
||||||
|
.send(Message::Text(
|
||||||
|
serde_json::to_string(&payload).map_err(|_| ())?,
|
||||||
|
))
|
||||||
|
.await
|
||||||
|
.map_err(|_| ())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn send_ws_error(socket: &mut WebSocket, message: &str) -> Result<(), ()> {
|
||||||
|
send_ws_json(
|
||||||
|
socket,
|
||||||
|
json!({
|
||||||
|
"type": "error",
|
||||||
|
"message": message,
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
|
||||||
#[utoipa::path(
|
#[utoipa::path(
|
||||||
get,
|
get,
|
||||||
path = "/v1/config/mcp",
|
path = "/v1/config/mcp",
|
||||||
|
|
@ -1386,6 +2134,96 @@ async fn delete_v1_acp(
|
||||||
Ok(StatusCode::NO_CONTENT)
|
Ok(StatusCode::NO_CONTENT)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn process_api_supported() -> bool {
|
||||||
|
!cfg!(windows)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn process_api_not_supported() -> ProblemDetails {
|
||||||
|
ProblemDetails {
|
||||||
|
type_: ErrorType::InvalidRequest.as_urn().to_string(),
|
||||||
|
title: "Not Implemented".to_string(),
|
||||||
|
status: 501,
|
||||||
|
detail: Some("process API is not implemented on Windows".to_string()),
|
||||||
|
instance: None,
|
||||||
|
extensions: serde_json::Map::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn map_process_config(config: ProcessRuntimeConfig) -> ProcessConfig {
|
||||||
|
ProcessConfig {
|
||||||
|
max_concurrent_processes: config.max_concurrent_processes,
|
||||||
|
default_run_timeout_ms: config.default_run_timeout_ms,
|
||||||
|
max_run_timeout_ms: config.max_run_timeout_ms,
|
||||||
|
max_output_bytes: config.max_output_bytes,
|
||||||
|
max_log_bytes_per_process: config.max_log_bytes_per_process,
|
||||||
|
max_input_bytes_per_request: config.max_input_bytes_per_request,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn into_runtime_process_config(config: ProcessConfig) -> ProcessRuntimeConfig {
|
||||||
|
ProcessRuntimeConfig {
|
||||||
|
max_concurrent_processes: config.max_concurrent_processes,
|
||||||
|
default_run_timeout_ms: config.default_run_timeout_ms,
|
||||||
|
max_run_timeout_ms: config.max_run_timeout_ms,
|
||||||
|
max_output_bytes: config.max_output_bytes,
|
||||||
|
max_log_bytes_per_process: config.max_log_bytes_per_process,
|
||||||
|
max_input_bytes_per_request: config.max_input_bytes_per_request,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn map_process_snapshot(snapshot: ProcessSnapshot) -> ProcessInfo {
|
||||||
|
ProcessInfo {
|
||||||
|
id: snapshot.id,
|
||||||
|
command: snapshot.command,
|
||||||
|
args: snapshot.args,
|
||||||
|
cwd: snapshot.cwd,
|
||||||
|
tty: snapshot.tty,
|
||||||
|
interactive: snapshot.interactive,
|
||||||
|
status: match snapshot.status {
|
||||||
|
ProcessStatus::Running => ProcessState::Running,
|
||||||
|
ProcessStatus::Exited => ProcessState::Exited,
|
||||||
|
},
|
||||||
|
pid: snapshot.pid,
|
||||||
|
exit_code: snapshot.exit_code,
|
||||||
|
created_at_ms: snapshot.created_at_ms,
|
||||||
|
exited_at_ms: snapshot.exited_at_ms,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn into_runtime_log_stream(stream: ProcessLogsStream) -> ProcessLogFilterStream {
|
||||||
|
match stream {
|
||||||
|
ProcessLogsStream::Stdout => ProcessLogFilterStream::Stdout,
|
||||||
|
ProcessLogsStream::Stderr => ProcessLogFilterStream::Stderr,
|
||||||
|
ProcessLogsStream::Combined => ProcessLogFilterStream::Combined,
|
||||||
|
ProcessLogsStream::Pty => ProcessLogFilterStream::Pty,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn map_process_log_line(line: crate::process_runtime::ProcessLogLine) -> ProcessLogEntry {
|
||||||
|
ProcessLogEntry {
|
||||||
|
sequence: line.sequence,
|
||||||
|
stream: match line.stream {
|
||||||
|
ProcessStream::Stdout => ProcessLogsStream::Stdout,
|
||||||
|
ProcessStream::Stderr => ProcessLogsStream::Stderr,
|
||||||
|
ProcessStream::Pty => ProcessLogsStream::Pty,
|
||||||
|
},
|
||||||
|
timestamp_ms: line.timestamp_ms,
|
||||||
|
data: line.data,
|
||||||
|
encoding: line.encoding.to_string(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn process_log_matches(entry: &ProcessLogEntry, stream: ProcessLogsStream) -> bool {
|
||||||
|
match stream {
|
||||||
|
ProcessLogsStream::Stdout => entry.stream == ProcessLogsStream::Stdout,
|
||||||
|
ProcessLogsStream::Stderr => entry.stream == ProcessLogsStream::Stderr,
|
||||||
|
ProcessLogsStream::Combined => {
|
||||||
|
entry.stream == ProcessLogsStream::Stdout || entry.stream == ProcessLogsStream::Stderr
|
||||||
|
}
|
||||||
|
ProcessLogsStream::Pty => entry.stream == ProcessLogsStream::Pty,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
fn validate_named_query(value: &str, field_name: &str) -> Result<(), SandboxError> {
|
fn validate_named_query(value: &str, field_name: &str) -> Result<(), SandboxError> {
|
||||||
if value.trim().is_empty() {
|
if value.trim().is_empty() {
|
||||||
return Err(SandboxError::InvalidRequest {
|
return Err(SandboxError::InvalidRequest {
|
||||||
|
|
|
||||||
|
|
@ -33,7 +33,17 @@ pub(super) async fn require_token(
|
||||||
.and_then(|value| value.to_str().ok())
|
.and_then(|value| value.to_str().ok())
|
||||||
.and_then(|value| value.strip_prefix("Bearer "));
|
.and_then(|value| value.strip_prefix("Bearer "));
|
||||||
|
|
||||||
if bearer == Some(expected.as_str()) {
|
let allow_query_token = request.uri().path().ends_with("/terminal/ws");
|
||||||
|
let query_token = if allow_query_token {
|
||||||
|
request
|
||||||
|
.uri()
|
||||||
|
.query()
|
||||||
|
.and_then(|query| query_param(query, "access_token"))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
};
|
||||||
|
|
||||||
|
if bearer == Some(expected.as_str()) || query_token.as_deref() == Some(expected.as_str()) {
|
||||||
return Ok(next.run(request).await);
|
return Ok(next.run(request).await);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -42,6 +52,53 @@ pub(super) async fn require_token(
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn query_param(query: &str, key: &str) -> Option<String> {
|
||||||
|
query
|
||||||
|
.split('&')
|
||||||
|
.filter_map(|part| part.split_once('='))
|
||||||
|
.find_map(|(k, v)| {
|
||||||
|
if k == key {
|
||||||
|
Some(percent_decode(v))
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
fn percent_decode(input: &str) -> String {
|
||||||
|
let mut output = Vec::with_capacity(input.len());
|
||||||
|
let bytes = input.as_bytes();
|
||||||
|
let mut i = 0;
|
||||||
|
while i < bytes.len() {
|
||||||
|
if bytes[i] == b'%' && i + 2 < bytes.len() {
|
||||||
|
if let (Some(hi), Some(lo)) = (
|
||||||
|
hex_nibble(bytes[i + 1]),
|
||||||
|
hex_nibble(bytes[i + 2]),
|
||||||
|
) {
|
||||||
|
output.push((hi << 4) | lo);
|
||||||
|
i += 3;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if bytes[i] == b'+' {
|
||||||
|
output.push(b' ');
|
||||||
|
} else {
|
||||||
|
output.push(bytes[i]);
|
||||||
|
}
|
||||||
|
i += 1;
|
||||||
|
}
|
||||||
|
String::from_utf8(output).unwrap_or_else(|_| input.to_string())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn hex_nibble(b: u8) -> Option<u8> {
|
||||||
|
match b {
|
||||||
|
b'0'..=b'9' => Some(b - b'0'),
|
||||||
|
b'a'..=b'f' => Some(b - b'a' + 10),
|
||||||
|
b'A'..=b'F' => Some(b - b'A' + 10),
|
||||||
|
_ => None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub(super) type PinBoxSseStream = crate::acp_proxy_runtime::PinBoxSseStream;
|
pub(super) type PinBoxSseStream = crate::acp_proxy_runtime::PinBoxSseStream;
|
||||||
|
|
||||||
pub(super) fn credentials_available_for(
|
pub(super) fn credentials_available_for(
|
||||||
|
|
@ -547,8 +604,17 @@ pub(super) fn problem_from_sandbox_error(error: &SandboxError) -> ProblemDetails
|
||||||
let mut problem = error.to_problem_details();
|
let mut problem = error.to_problem_details();
|
||||||
|
|
||||||
match error {
|
match error {
|
||||||
SandboxError::InvalidRequest { .. } => {
|
SandboxError::InvalidRequest { message } => {
|
||||||
problem.status = 400;
|
if message.starts_with("input payload exceeds maxInputBytesPerRequest") {
|
||||||
|
problem.status = 413;
|
||||||
|
problem.title = "Payload Too Large".to_string();
|
||||||
|
} else {
|
||||||
|
problem.status = 400;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
SandboxError::NotFound { .. } => {
|
||||||
|
problem.status = 404;
|
||||||
|
problem.title = "Not Found".to_string();
|
||||||
}
|
}
|
||||||
SandboxError::Timeout { .. } => {
|
SandboxError::Timeout { .. } => {
|
||||||
problem.status = 504;
|
problem.status = 504;
|
||||||
|
|
|
||||||
|
|
@ -362,3 +362,173 @@ pub struct AcpEnvelope {
|
||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
pub error: Option<Value>,
|
pub error: Option<Value>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessConfig {
|
||||||
|
pub max_concurrent_processes: usize,
|
||||||
|
pub default_run_timeout_ms: u64,
|
||||||
|
pub max_run_timeout_ms: u64,
|
||||||
|
pub max_output_bytes: usize,
|
||||||
|
pub max_log_bytes_per_process: usize,
|
||||||
|
pub max_input_bytes_per_request: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessCreateRequest {
|
||||||
|
pub command: String,
|
||||||
|
#[serde(default)]
|
||||||
|
pub args: Vec<String>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub cwd: Option<String>,
|
||||||
|
#[serde(default, skip_serializing_if = "BTreeMap::is_empty")]
|
||||||
|
pub env: BTreeMap<String, String>,
|
||||||
|
#[serde(default)]
|
||||||
|
pub tty: bool,
|
||||||
|
#[serde(default)]
|
||||||
|
pub interactive: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessRunRequest {
|
||||||
|
pub command: String,
|
||||||
|
#[serde(default)]
|
||||||
|
pub args: Vec<String>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub cwd: Option<String>,
|
||||||
|
#[serde(default, skip_serializing_if = "BTreeMap::is_empty")]
|
||||||
|
pub env: BTreeMap<String, String>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub timeout_ms: Option<u64>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub max_output_bytes: Option<usize>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessRunResponse {
|
||||||
|
pub exit_code: Option<i32>,
|
||||||
|
pub timed_out: bool,
|
||||||
|
pub stdout: String,
|
||||||
|
pub stderr: String,
|
||||||
|
pub stdout_truncated: bool,
|
||||||
|
pub stderr_truncated: bool,
|
||||||
|
pub duration_ms: u64,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema, PartialEq, Eq)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
pub enum ProcessState {
|
||||||
|
Running,
|
||||||
|
Exited,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessInfo {
|
||||||
|
pub id: String,
|
||||||
|
pub command: String,
|
||||||
|
pub args: Vec<String>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub cwd: Option<String>,
|
||||||
|
pub tty: bool,
|
||||||
|
pub interactive: bool,
|
||||||
|
pub status: ProcessState,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub pid: Option<u32>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub exit_code: Option<i32>,
|
||||||
|
pub created_at_ms: i64,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub exited_at_ms: Option<i64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessListResponse {
|
||||||
|
pub processes: Vec<ProcessInfo>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Copy, Serialize, Deserialize, JsonSchema, ToSchema, PartialEq, Eq)]
|
||||||
|
#[serde(rename_all = "lowercase")]
|
||||||
|
pub enum ProcessLogsStream {
|
||||||
|
Stdout,
|
||||||
|
Stderr,
|
||||||
|
Combined,
|
||||||
|
Pty,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessLogsQuery {
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub stream: Option<ProcessLogsStream>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub tail: Option<usize>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub follow: Option<bool>,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub since: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessLogEntry {
|
||||||
|
pub sequence: u64,
|
||||||
|
pub stream: ProcessLogsStream,
|
||||||
|
pub timestamp_ms: i64,
|
||||||
|
pub data: String,
|
||||||
|
pub encoding: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessLogsResponse {
|
||||||
|
pub process_id: String,
|
||||||
|
pub stream: ProcessLogsStream,
|
||||||
|
pub entries: Vec<ProcessLogEntry>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessInputRequest {
|
||||||
|
pub data: String,
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub encoding: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessInputResponse {
|
||||||
|
pub bytes_written: usize,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessSignalQuery {
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub wait_ms: Option<u64>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessTerminalResizeRequest {
|
||||||
|
pub cols: u16,
|
||||||
|
pub rows: u16,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Serialize, Deserialize, JsonSchema, ToSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessTerminalResizeResponse {
|
||||||
|
pub cols: u16,
|
||||||
|
pub rows: u16,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, Clone, Deserialize, JsonSchema)]
|
||||||
|
#[serde(rename_all = "camelCase")]
|
||||||
|
pub struct ProcessWsQuery {
|
||||||
|
#[serde(default, skip_serializing_if = "Option::is_none")]
|
||||||
|
pub access_token: Option<String>,
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,6 @@
|
||||||
use std::fs;
|
use std::fs;
|
||||||
use std::io::{Read, Write};
|
use std::io::{Read, Write};
|
||||||
use std::net::{TcpListener, TcpStream};
|
use std::net::{SocketAddr, TcpListener, TcpStream};
|
||||||
use std::path::Path;
|
use std::path::Path;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
|
|
||||||
|
|
@ -14,6 +14,8 @@ use sandbox_agent_agent_management::agents::AgentManager;
|
||||||
use serde_json::{json, Value};
|
use serde_json::{json, Value};
|
||||||
use serial_test::serial;
|
use serial_test::serial;
|
||||||
use tempfile::TempDir;
|
use tempfile::TempDir;
|
||||||
|
use tokio::sync::oneshot;
|
||||||
|
use tokio::task::JoinHandle;
|
||||||
use tower::util::ServiceExt;
|
use tower::util::ServiceExt;
|
||||||
|
|
||||||
struct TestApp {
|
struct TestApp {
|
||||||
|
|
@ -48,6 +50,56 @@ struct EnvVarGuard {
|
||||||
previous: Option<std::ffi::OsString>,
|
previous: Option<std::ffi::OsString>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct LiveServer {
|
||||||
|
address: SocketAddr,
|
||||||
|
shutdown_tx: Option<oneshot::Sender<()>>,
|
||||||
|
task: JoinHandle<()>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl LiveServer {
|
||||||
|
async fn spawn(app: Router) -> Self {
|
||||||
|
let listener = tokio::net::TcpListener::bind("127.0.0.1:0")
|
||||||
|
.await
|
||||||
|
.expect("bind live server");
|
||||||
|
let address = listener.local_addr().expect("live server address");
|
||||||
|
let (shutdown_tx, shutdown_rx) = oneshot::channel::<()>();
|
||||||
|
|
||||||
|
let task = tokio::spawn(async move {
|
||||||
|
let server = axum::serve(listener, app.into_make_service())
|
||||||
|
.with_graceful_shutdown(async {
|
||||||
|
let _ = shutdown_rx.await;
|
||||||
|
});
|
||||||
|
|
||||||
|
let _ = server.await;
|
||||||
|
});
|
||||||
|
|
||||||
|
Self {
|
||||||
|
address,
|
||||||
|
shutdown_tx: Some(shutdown_tx),
|
||||||
|
task,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn http_url(&self, path: &str) -> String {
|
||||||
|
format!("http://{}{}", self.address, path)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn ws_url(&self, path: &str) -> String {
|
||||||
|
format!("ws://{}{}", self.address, path)
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn shutdown(mut self) {
|
||||||
|
if let Some(shutdown_tx) = self.shutdown_tx.take() {
|
||||||
|
let _ = shutdown_tx.send(());
|
||||||
|
}
|
||||||
|
|
||||||
|
let _ = tokio::time::timeout(Duration::from_secs(3), async {
|
||||||
|
let _ = self.task.await;
|
||||||
|
})
|
||||||
|
.await;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl EnvVarGuard {
|
impl EnvVarGuard {
|
||||||
fn set(key: &'static str, value: &str) -> Self {
|
fn set(key: &'static str, value: &str) -> Self {
|
||||||
let previous = std::env::var_os(key);
|
let previous = std::env::var_os(key);
|
||||||
|
|
@ -291,3 +343,5 @@ mod acp_transport;
|
||||||
mod config_endpoints;
|
mod config_endpoints;
|
||||||
#[path = "v1_api/control_plane.rs"]
|
#[path = "v1_api/control_plane.rs"]
|
||||||
mod control_plane;
|
mod control_plane;
|
||||||
|
#[path = "v1_api/processes.rs"]
|
||||||
|
mod processes;
|
||||||
|
|
|
||||||
661
server/packages/sandbox-agent/tests/v1_api/processes.rs
Normal file
661
server/packages/sandbox-agent/tests/v1_api/processes.rs
Normal file
|
|
@ -0,0 +1,661 @@
|
||||||
|
use super::*;
|
||||||
|
use base64::engine::general_purpose::STANDARD as BASE64;
|
||||||
|
use base64::Engine;
|
||||||
|
use futures::{SinkExt, StreamExt};
|
||||||
|
use tokio_tungstenite::connect_async;
|
||||||
|
use tokio_tungstenite::tungstenite::Message;
|
||||||
|
|
||||||
|
async fn wait_for_exited(test_app: &TestApp, process_id: &str) {
|
||||||
|
for _ in 0..30 {
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
&format!("/v1/processes/{process_id}"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let parsed = parse_json(&body);
|
||||||
|
if parsed["status"] == "exited" {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
tokio::time::sleep(Duration::from_millis(100)).await;
|
||||||
|
}
|
||||||
|
|
||||||
|
panic!("process did not exit in time");
|
||||||
|
}
|
||||||
|
|
||||||
|
fn decode_log_entries(entries: &[Value]) -> String {
|
||||||
|
entries
|
||||||
|
.iter()
|
||||||
|
.filter_map(|entry| entry.get("data").and_then(Value::as_str))
|
||||||
|
.filter_map(|encoded| BASE64.decode(encoded).ok())
|
||||||
|
.map(|bytes| String::from_utf8_lossy(&bytes).to_string())
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
.join("")
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn recv_ws_message(
|
||||||
|
ws: &mut tokio_tungstenite::WebSocketStream<
|
||||||
|
tokio_tungstenite::MaybeTlsStream<tokio::net::TcpStream>,
|
||||||
|
>,
|
||||||
|
) -> Message {
|
||||||
|
tokio::time::timeout(Duration::from_secs(3), ws.next())
|
||||||
|
.await
|
||||||
|
.expect("timed out waiting for websocket frame")
|
||||||
|
.expect("websocket stream ended")
|
||||||
|
.expect("websocket frame")
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_processes_config_round_trip() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
"/v1/processes/config",
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
assert_eq!(parse_json(&body)["maxConcurrentProcesses"], 64);
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes/config",
|
||||||
|
Some(json!({
|
||||||
|
"maxConcurrentProcesses": 8,
|
||||||
|
"defaultRunTimeoutMs": 1000,
|
||||||
|
"maxRunTimeoutMs": 5000,
|
||||||
|
"maxOutputBytes": 4096,
|
||||||
|
"maxLogBytesPerProcess": 32768,
|
||||||
|
"maxInputBytesPerRequest": 1024
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let parsed = parse_json(&body);
|
||||||
|
assert_eq!(parsed["maxConcurrentProcesses"], 8);
|
||||||
|
assert_eq!(parsed["defaultRunTimeoutMs"], 1000);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_lifecycle_requires_stop_before_delete() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "sleep 30"],
|
||||||
|
"tty": false,
|
||||||
|
"interactive": false
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::DELETE,
|
||||||
|
&format!("/v1/processes/{process_id}"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::CONFLICT);
|
||||||
|
assert_eq!(parse_json(&body)["status"], 409);
|
||||||
|
|
||||||
|
let (status, _, _body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
&format!("/v1/processes/{process_id}/stop"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
|
||||||
|
wait_for_exited(&test_app, &process_id).await;
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::DELETE,
|
||||||
|
&format!("/v1/processes/{process_id}"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::NO_CONTENT);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_run_returns_output_and_timeout() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes/run",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "echo hi"],
|
||||||
|
"timeoutMs": 1000
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let parsed = parse_json(&body);
|
||||||
|
assert_eq!(parsed["timedOut"], false);
|
||||||
|
assert_eq!(parsed["exitCode"], 0);
|
||||||
|
assert!(parsed["stdout"].as_str().unwrap_or_default().contains("hi"));
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes/run",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "sleep 2"],
|
||||||
|
"timeoutMs": 50
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
assert_eq!(parse_json(&body)["timedOut"], true);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_run_reports_truncation() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes/run",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "printf 'abcdefghijklmnopqrstuvwxyz'"],
|
||||||
|
"maxOutputBytes": 5
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let parsed = parse_json(&body);
|
||||||
|
assert_eq!(parsed["stdoutTruncated"], true);
|
||||||
|
assert_eq!(parsed["stderrTruncated"], false);
|
||||||
|
assert_eq!(parsed["stdout"].as_str().unwrap_or_default().len(), 5);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_tty_input_and_logs() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "cat",
|
||||||
|
"tty": true,
|
||||||
|
"interactive": true
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let (status, _, _body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
&format!("/v1/processes/{process_id}/input"),
|
||||||
|
Some(json!({
|
||||||
|
"data": "aGVsbG8K",
|
||||||
|
"encoding": "base64"
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
|
||||||
|
tokio::time::sleep(Duration::from_millis(150)).await;
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
&format!("/v1/processes/{process_id}/logs?stream=pty&tail=20"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let entries = parse_json(&body)["entries"]
|
||||||
|
.as_array()
|
||||||
|
.cloned()
|
||||||
|
.unwrap_or_default();
|
||||||
|
assert!(!entries.is_empty());
|
||||||
|
|
||||||
|
let (status, _, _body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
&format!("/v1/processes/{process_id}/kill"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
|
||||||
|
wait_for_exited(&test_app, &process_id).await;
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::DELETE,
|
||||||
|
&format!("/v1/processes/{process_id}"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::NO_CONTENT);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_not_found_returns_404() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
"/v1/processes/does-not-exist",
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::NOT_FOUND);
|
||||||
|
assert_eq!(parse_json(&body)["status"], 404);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_input_limit_returns_413() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes/config",
|
||||||
|
Some(json!({
|
||||||
|
"maxConcurrentProcesses": 8,
|
||||||
|
"defaultRunTimeoutMs": 1000,
|
||||||
|
"maxRunTimeoutMs": 5000,
|
||||||
|
"maxOutputBytes": 4096,
|
||||||
|
"maxLogBytesPerProcess": 32768,
|
||||||
|
"maxInputBytesPerRequest": 4
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "cat",
|
||||||
|
"tty": true,
|
||||||
|
"interactive": true
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
&format!("/v1/processes/{process_id}/input"),
|
||||||
|
Some(json!({
|
||||||
|
"data": "aGVsbG8=",
|
||||||
|
"encoding": "base64"
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::PAYLOAD_TOO_LARGE);
|
||||||
|
assert_eq!(parse_json(&body)["status"], 413);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_tty_process_is_real_terminal() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "tty"],
|
||||||
|
"tty": true,
|
||||||
|
"interactive": false
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
wait_for_exited(&test_app, &process_id).await;
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
&format!("/v1/processes/{process_id}/logs?stream=pty"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let entries = parse_json(&body)["entries"]
|
||||||
|
.as_array()
|
||||||
|
.cloned()
|
||||||
|
.unwrap_or_default();
|
||||||
|
let joined = decode_log_entries(&entries);
|
||||||
|
assert!(!joined.to_lowercase().contains("not a tty"));
|
||||||
|
assert!(joined.contains("/dev/"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_logs_follow_sse_streams_entries() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "echo first; sleep 0.3; echo second"],
|
||||||
|
"tty": false,
|
||||||
|
"interactive": false
|
||||||
|
})),
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let request = Request::builder()
|
||||||
|
.method(Method::GET)
|
||||||
|
.uri(format!(
|
||||||
|
"/v1/processes/{process_id}/logs?stream=stdout&follow=true"
|
||||||
|
))
|
||||||
|
.body(Body::empty())
|
||||||
|
.expect("build request");
|
||||||
|
let response = test_app
|
||||||
|
.app
|
||||||
|
.clone()
|
||||||
|
.oneshot(request)
|
||||||
|
.await
|
||||||
|
.expect("sse response");
|
||||||
|
assert_eq!(response.status(), StatusCode::OK);
|
||||||
|
|
||||||
|
let mut stream = response.into_body().into_data_stream();
|
||||||
|
let chunk = tokio::time::timeout(Duration::from_secs(5), async move {
|
||||||
|
while let Some(chunk) = stream.next().await {
|
||||||
|
let bytes = chunk.expect("stream chunk");
|
||||||
|
let text = String::from_utf8_lossy(&bytes).to_string();
|
||||||
|
if text.contains("data:") {
|
||||||
|
return text;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
panic!("SSE stream ended before log chunk");
|
||||||
|
})
|
||||||
|
.await
|
||||||
|
.expect("timed out reading process log sse");
|
||||||
|
|
||||||
|
let payload = parse_sse_data(&chunk);
|
||||||
|
assert!(payload["sequence"].as_u64().is_some());
|
||||||
|
assert_eq!(payload["stream"], "stdout");
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_access_token_query_only_allows_terminal_ws() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::with_token("secret-token".to_string()));
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
"/v1/health?access_token=secret-token",
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::UNAUTHORIZED);
|
||||||
|
|
||||||
|
let (status, _, body) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::POST,
|
||||||
|
"/v1/processes",
|
||||||
|
Some(json!({
|
||||||
|
"command": "cat",
|
||||||
|
"tty": true,
|
||||||
|
"interactive": true
|
||||||
|
})),
|
||||||
|
&[("authorization", "Bearer secret-token")],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::OK);
|
||||||
|
let process_id = parse_json(&body)["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
&format!("/v1/processes/{process_id}/terminal/ws"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::UNAUTHORIZED);
|
||||||
|
|
||||||
|
let (status, _, _) = send_request(
|
||||||
|
&test_app.app,
|
||||||
|
Method::GET,
|
||||||
|
&format!("/v1/processes/{process_id}/terminal/ws?access_token=secret-token"),
|
||||||
|
None,
|
||||||
|
&[],
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
assert_eq!(status, StatusCode::BAD_REQUEST);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_terminal_ws_e2e_is_deterministic() {
|
||||||
|
let test_app = TestApp::new(AuthConfig::disabled());
|
||||||
|
let live_server = LiveServer::spawn(test_app.app.clone()).await;
|
||||||
|
let http = reqwest::Client::new();
|
||||||
|
|
||||||
|
let create_response = http
|
||||||
|
.post(live_server.http_url("/v1/processes"))
|
||||||
|
.json(&json!({
|
||||||
|
"command": "sh",
|
||||||
|
"args": ["-lc", "stty -echo; IFS= read -r line; printf 'got:%s\\n' \"$line\""],
|
||||||
|
"tty": true,
|
||||||
|
"interactive": true
|
||||||
|
}))
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.expect("create process response");
|
||||||
|
assert_eq!(create_response.status(), reqwest::StatusCode::OK);
|
||||||
|
let create_body: Value = create_response.json().await.expect("create process json");
|
||||||
|
let process_id = create_body["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let ws_url = live_server.ws_url(&format!("/v1/processes/{process_id}/terminal/ws"));
|
||||||
|
let (mut ws, _) = connect_async(&ws_url)
|
||||||
|
.await
|
||||||
|
.expect("connect websocket");
|
||||||
|
|
||||||
|
let ready = recv_ws_message(&mut ws).await;
|
||||||
|
let ready_payload: Value = serde_json::from_str(ready.to_text().expect("ready text frame"))
|
||||||
|
.expect("ready json");
|
||||||
|
assert_eq!(ready_payload["type"], "ready");
|
||||||
|
assert_eq!(ready_payload["processId"], process_id);
|
||||||
|
|
||||||
|
ws.send(Message::Text(
|
||||||
|
json!({
|
||||||
|
"type": "input",
|
||||||
|
"data": "hello from ws\n"
|
||||||
|
})
|
||||||
|
.to_string(),
|
||||||
|
))
|
||||||
|
.await
|
||||||
|
.expect("send input frame");
|
||||||
|
|
||||||
|
let mut saw_binary_output = false;
|
||||||
|
let mut saw_exit = false;
|
||||||
|
for _ in 0..10 {
|
||||||
|
let frame = recv_ws_message(&mut ws).await;
|
||||||
|
match frame {
|
||||||
|
Message::Binary(bytes) => {
|
||||||
|
let text = String::from_utf8_lossy(&bytes);
|
||||||
|
if text.contains("got:hello from ws") {
|
||||||
|
saw_binary_output = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Message::Text(text) => {
|
||||||
|
let payload: Value = serde_json::from_str(&text).expect("ws json");
|
||||||
|
if payload["type"] == "exit" {
|
||||||
|
saw_exit = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
assert_ne!(payload["type"], "error");
|
||||||
|
}
|
||||||
|
Message::Close(_) => break,
|
||||||
|
Message::Ping(_) | Message::Pong(_) => {}
|
||||||
|
_ => {}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
assert!(saw_binary_output, "expected pty binary output over websocket");
|
||||||
|
assert!(saw_exit, "expected exit control frame over websocket");
|
||||||
|
|
||||||
|
let _ = ws.close(None).await;
|
||||||
|
|
||||||
|
let delete_response = http
|
||||||
|
.delete(live_server.http_url(&format!("/v1/processes/{process_id}")))
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.expect("delete process response");
|
||||||
|
assert_eq!(delete_response.status(), reqwest::StatusCode::NO_CONTENT);
|
||||||
|
|
||||||
|
live_server.shutdown().await;
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn v1_process_terminal_ws_auth_e2e() {
|
||||||
|
let token = "secret-token";
|
||||||
|
let test_app = TestApp::new(AuthConfig::with_token(token.to_string()));
|
||||||
|
let live_server = LiveServer::spawn(test_app.app.clone()).await;
|
||||||
|
let http = reqwest::Client::new();
|
||||||
|
|
||||||
|
let create_response = http
|
||||||
|
.post(live_server.http_url("/v1/processes"))
|
||||||
|
.bearer_auth(token)
|
||||||
|
.json(&json!({
|
||||||
|
"command": "cat",
|
||||||
|
"tty": true,
|
||||||
|
"interactive": true
|
||||||
|
}))
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.expect("create process response");
|
||||||
|
assert_eq!(create_response.status(), reqwest::StatusCode::OK);
|
||||||
|
let create_body: Value = create_response.json().await.expect("create process json");
|
||||||
|
let process_id = create_body["id"]
|
||||||
|
.as_str()
|
||||||
|
.expect("process id")
|
||||||
|
.to_string();
|
||||||
|
|
||||||
|
let unauth_ws_url = live_server.ws_url(&format!("/v1/processes/{process_id}/terminal/ws"));
|
||||||
|
let unauth_err = connect_async(&unauth_ws_url)
|
||||||
|
.await
|
||||||
|
.expect_err("unauthenticated websocket handshake should fail");
|
||||||
|
match unauth_err {
|
||||||
|
tokio_tungstenite::tungstenite::Error::Http(response) => {
|
||||||
|
assert_eq!(response.status().as_u16(), 401);
|
||||||
|
}
|
||||||
|
other => panic!("unexpected websocket auth error: {other:?}"),
|
||||||
|
}
|
||||||
|
|
||||||
|
let auth_ws_url = live_server.ws_url(&format!(
|
||||||
|
"/v1/processes/{process_id}/terminal/ws?access_token={token}"
|
||||||
|
));
|
||||||
|
let (mut ws, _) = connect_async(&auth_ws_url)
|
||||||
|
.await
|
||||||
|
.expect("authenticated websocket handshake");
|
||||||
|
|
||||||
|
let ready = recv_ws_message(&mut ws).await;
|
||||||
|
let ready_payload: Value = serde_json::from_str(ready.to_text().expect("ready text frame"))
|
||||||
|
.expect("ready json");
|
||||||
|
assert_eq!(ready_payload["type"], "ready");
|
||||||
|
assert_eq!(ready_payload["processId"], process_id);
|
||||||
|
|
||||||
|
let _ = ws
|
||||||
|
.send(Message::Text(json!({ "type": "close" }).to_string()))
|
||||||
|
.await;
|
||||||
|
let _ = ws.close(None).await;
|
||||||
|
|
||||||
|
let kill_response = http
|
||||||
|
.post(live_server.http_url(&format!(
|
||||||
|
"/v1/processes/{process_id}/kill?waitMs=1000"
|
||||||
|
)))
|
||||||
|
.bearer_auth(token)
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.expect("kill process response");
|
||||||
|
assert_eq!(kill_response.status(), reqwest::StatusCode::OK);
|
||||||
|
|
||||||
|
let delete_response = http
|
||||||
|
.delete(live_server.http_url(&format!("/v1/processes/{process_id}")))
|
||||||
|
.bearer_auth(token)
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.expect("delete process response");
|
||||||
|
assert_eq!(delete_response.status(), reqwest::StatusCode::NO_CONTENT);
|
||||||
|
|
||||||
|
live_server.shutdown().await;
|
||||||
|
}
|
||||||
Loading…
Add table
Add a link
Reference in a new issue