Files
openclaw/src/media-understanding/runtime.ts
Daniel Alkurdi 1f0431cd11 fix(media): surface OpenAI audio transcription failures (#65096)
* fix(media): surface audio transcription provider failures

* fix(media): prefer failed reasons in surfaced errors

* fix(media): import attempt outcome type

* fix(media): guard malformed decision arrays

---------

Co-authored-by: Vincent Koc <vincentkoc@ieee.org>
2026-04-12 17:05:18 +01:00

165 lines
5.1 KiB
TypeScript

import fs from "node:fs/promises";
import path from "node:path";
import { normalizeMediaProviderId } from "./provider-registry.js";
import { findDecisionReason, normalizeDecisionReason } from "./runner.entries.js";
import {
buildProviderRegistry,
createMediaAttachmentCache,
normalizeMediaAttachments,
runCapability,
} from "./runner.js";
import type {
DescribeImageFileParams,
DescribeImageFileWithModelParams,
DescribeVideoFileParams,
RunMediaUnderstandingFileParams,
RunMediaUnderstandingFileResult,
TranscribeAudioFileParams,
} from "./runtime-types.js";
export type {
DescribeImageFileParams,
DescribeImageFileWithModelParams,
DescribeVideoFileParams,
RunMediaUnderstandingFileParams,
RunMediaUnderstandingFileResult,
TranscribeAudioFileParams,
} from "./runtime-types.js";
type MediaUnderstandingCapability = "image" | "audio" | "video";
type MediaUnderstandingOutput = Awaited<ReturnType<typeof runCapability>>["outputs"][number];
const KIND_BY_CAPABILITY: Record<MediaUnderstandingCapability, MediaUnderstandingOutput["kind"]> = {
audio: "audio.transcription",
image: "image.description",
video: "video.description",
};
function resolveDecisionFailureReason(
decision: Awaited<ReturnType<typeof runCapability>>["decision"],
): string | undefined {
return normalizeDecisionReason(findDecisionReason(decision, "failed"));
}
function buildFileContext(params: { filePath: string; mime?: string }) {
return {
MediaPath: params.filePath,
MediaType: params.mime,
};
}
export async function runMediaUnderstandingFile(
params: RunMediaUnderstandingFileParams,
): Promise<RunMediaUnderstandingFileResult> {
const ctx = buildFileContext(params);
const attachments = normalizeMediaAttachments(ctx);
if (attachments.length === 0) {
return { text: undefined };
}
const config = params.cfg.tools?.media?.[params.capability];
if (config?.enabled === false) {
return {
text: undefined,
provider: undefined,
model: undefined,
output: undefined,
};
}
const providerRegistry = buildProviderRegistry(undefined, params.cfg);
const cache = createMediaAttachmentCache(attachments, {
localPathRoots: [path.dirname(params.filePath)],
});
try {
const result = await runCapability({
capability: params.capability,
cfg: params.cfg,
ctx,
attachments: cache,
media: attachments,
agentDir: params.agentDir,
providerRegistry,
config,
activeModel: params.activeModel,
});
if (result.outputs.length === 0 && result.decision.outcome === "failed") {
throw new Error(
resolveDecisionFailureReason(result.decision) ??
`${params.capability} understanding failed`,
);
}
const output = result.outputs.find(
(entry) => entry.kind === KIND_BY_CAPABILITY[params.capability],
);
const text = output?.text?.trim();
return {
text: text || undefined,
provider: output?.provider,
model: output?.model,
output,
};
} finally {
await cache.cleanup();
}
}
export async function describeImageFile(
params: DescribeImageFileParams,
): Promise<RunMediaUnderstandingFileResult> {
return await runMediaUnderstandingFile({ ...params, capability: "image" });
}
export async function describeImageFileWithModel(params: DescribeImageFileWithModelParams) {
const timeoutMs = params.timeoutMs ?? 30_000;
const providerRegistry = buildProviderRegistry(undefined, params.cfg);
const provider = providerRegistry.get(normalizeMediaProviderId(params.provider));
if (!provider?.describeImage) {
throw new Error(`Provider does not support image analysis: ${params.provider}`);
}
const buffer = await fs.readFile(params.filePath);
return await provider.describeImage({
buffer,
fileName: path.basename(params.filePath),
mime: params.mime,
provider: params.provider,
model: params.model,
prompt: params.prompt,
maxTokens: params.maxTokens,
timeoutMs,
cfg: params.cfg,
agentDir: params.agentDir ?? "",
});
}
export async function describeVideoFile(
params: DescribeVideoFileParams,
): Promise<RunMediaUnderstandingFileResult> {
return await runMediaUnderstandingFile({ ...params, capability: "video" });
}
export async function transcribeAudioFile(
params: TranscribeAudioFileParams,
): Promise<{ text: string | undefined }> {
const cfg =
params.language || params.prompt
? {
...params.cfg,
tools: {
...params.cfg.tools,
media: {
...params.cfg.tools?.media,
audio: {
...params.cfg.tools?.media?.audio,
...(params.language ? { _requestLanguageOverride: params.language } : {}),
...(params.prompt ? { _requestPromptOverride: params.prompt } : {}),
...(params.language ? { language: params.language } : {}),
...(params.prompt ? { prompt: params.prompt } : {}),
},
},
},
}
: params.cfg;
const result = await runMediaUnderstandingFile({ ...params, cfg, capability: "audio" });
return { text: result.text };
}