UNPKG

@axflow/models

Version:

Zero-dependency, modular SDK for building robust natural language applications

107 lines (104 loc) 3.08 kB
// src/openai/completion.ts import { POST, HttpError } from "@axflow/models/shared"; // src/openai/shared.ts function headers(apiKey, customHeaders) { const headers2 = { accept: "application/json", "content-type": "application/json", ...customHeaders }; if (typeof apiKey === "string") { headers2.authorization = `Bearer ${apiKey}`; } return headers2; } function streamTransformer(map) { let buffer = []; const decoder = new TextDecoder(); return (bytes, controller) => { const chunk = decoder.decode(bytes); for (let i = 0, len = chunk.length; i < len; ++i) { const isChunkSeparator = chunk[i] === "\n" && buffer[buffer.length - 1] === "\n"; if (!isChunkSeparator) { buffer.push(chunk[i]); continue; } const parsedChunk = parseChunk(buffer.join("")); if (parsedChunk) { controller.enqueue(map(parsedChunk)); } buffer = []; } }; } var DATA_RE = /data:\s*(.+)/; function parseChunk(chunk) { chunk = chunk.trim(); if (chunk.length === 0) { return null; } const match = chunk.match(DATA_RE); try { const data = match[1]; return data === "[DONE]" ? null : JSON.parse(data); } catch (error) { throw new Error( `Encountered unexpected chunk while parsing OpenAI streaming response: ${JSON.stringify( chunk )}` ); } } // src/openai/completion.ts var OPENAI_COMPLETIONS_API_URL = "https://api.openai.com/v1/completions"; async function run(request, options) { const url = options.apiUrl || OPENAI_COMPLETIONS_API_URL; const response = await POST(url, { headers: headers(options.apiKey, options.headers), body: JSON.stringify({ ...request, stream: false }), fetch: options.fetch, signal: options.signal }); return response.json(); } async function streamBytes(request, options) { const url = options.apiUrl || OPENAI_COMPLETIONS_API_URL; const response = await POST(url, { headers: headers(options.apiKey, options.headers), body: JSON.stringify({ ...request, stream: true }), fetch: options.fetch, signal: options.signal }); if (!response.body) { throw new HttpError("Expected response body to be a ReadableStream", response); } return response.body; } function noop(chunk) { return chunk; } async function stream(request, options) { const byteStream = await streamBytes(request, options); return byteStream.pipeThrough(new OpenAICompletionDecoderStream(noop)); } function chunkToToken(chunk) { return chunk.choices[0].text || ""; } async function streamTokens(request, options) { const byteStream = await streamBytes(request, options); return byteStream.pipeThrough(new OpenAICompletionDecoderStream(chunkToToken)); } var OpenAICompletion = class { static run = run; static stream = stream; static streamBytes = streamBytes; static streamTokens = streamTokens; }; var OpenAICompletionDecoderStream = class extends TransformStream { constructor(map) { super({ transform: streamTransformer(map) }); } }; export { OpenAICompletion };