mirror of
https://github.com/moltbot/moltbot.git
synced 2026-04-26 16:06:16 +00:00
* refactor(providers): centralize request attribution policy * style(providers): normalize request policy formatting * style(providers): normalize request policy formatting * style(providers): normalize request policy formatting * docs(changelog): note provider request policy fix * fix(providers): tighten request policy gates
35 lines
1.1 KiB
TypeScript
35 lines
1.1 KiB
TypeScript
import {
|
|
describeImageWithModel,
|
|
describeImagesWithModel,
|
|
transcribeOpenAiCompatibleAudio,
|
|
type AudioTranscriptionRequest,
|
|
type MediaUnderstandingProvider,
|
|
} from "openclaw/plugin-sdk/media-understanding";
|
|
import { OPENAI_DEFAULT_AUDIO_TRANSCRIPTION_MODEL } from "./default-models.js";
|
|
|
|
export const DEFAULT_OPENAI_AUDIO_BASE_URL = "https://api.openai.com/v1";
|
|
|
|
export async function transcribeOpenAiAudio(params: AudioTranscriptionRequest) {
|
|
return await transcribeOpenAiCompatibleAudio({
|
|
...params,
|
|
provider: "openai",
|
|
defaultBaseUrl: DEFAULT_OPENAI_AUDIO_BASE_URL,
|
|
defaultModel: OPENAI_DEFAULT_AUDIO_TRANSCRIPTION_MODEL,
|
|
});
|
|
}
|
|
|
|
export const openaiMediaUnderstandingProvider: MediaUnderstandingProvider = {
|
|
id: "openai",
|
|
capabilities: ["image", "audio"],
|
|
describeImage: describeImageWithModel,
|
|
describeImages: describeImagesWithModel,
|
|
transcribeAudio: transcribeOpenAiAudio,
|
|
};
|
|
|
|
export const openaiCodexMediaUnderstandingProvider: MediaUnderstandingProvider = {
|
|
id: "openai-codex",
|
|
capabilities: ["image"],
|
|
describeImage: describeImageWithModel,
|
|
describeImages: describeImagesWithModel,
|
|
};
|