UNPKG

@langchain/openai

Version:
42 lines (41 loc) 2.74 kB
import { BaseChatOpenAI, BaseChatOpenAICallOptions } from "./base.cjs"; import { OpenAI as OpenAI$1 } from "openai"; import { CallbackManagerForLLMRun } from "@langchain/core/callbacks/manager"; import { BaseMessage, BaseMessageChunk } from "@langchain/core/messages"; import { ChatGenerationChunk, ChatResult } from "@langchain/core/outputs"; //#region src/chat_models/completions.d.ts interface ChatOpenAICompletionsCallOptions extends BaseChatOpenAICallOptions {} type ChatCompletionsInvocationParams = Omit<OpenAI$1.Chat.Completions.ChatCompletionCreateParams, "messages">; /** * OpenAI Completions API implementation. * @internal */ declare class ChatOpenAICompletions<CallOptions extends ChatOpenAICompletionsCallOptions = ChatOpenAICompletionsCallOptions> extends BaseChatOpenAI<CallOptions> { /** @internal */ invocationParams(options?: this["ParsedCallOptions"], extra?: { streaming?: boolean; }): ChatCompletionsInvocationParams; _generate(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): Promise<ChatResult>; _streamResponseChunks(messages: BaseMessage[], options: this["ParsedCallOptions"], runManager?: CallbackManagerForLLMRun): AsyncGenerator<ChatGenerationChunk>; completionWithRetry(request: OpenAI$1.Chat.ChatCompletionCreateParamsStreaming, requestOptions?: OpenAI$1.RequestOptions): Promise<AsyncIterable<OpenAI$1.Chat.Completions.ChatCompletionChunk>>; completionWithRetry(request: OpenAI$1.Chat.ChatCompletionCreateParamsNonStreaming, requestOptions?: OpenAI$1.RequestOptions): Promise<OpenAI$1.Chat.Completions.ChatCompletion>; /** * @deprecated * This function was hoisted into a publicly accessible function from a * different export, but to maintain backwards compatibility with chat models * that depend on ChatOpenAICompletions, we'll keep it here as an overridable * method. This will be removed in a future release */ protected _convertCompletionsDeltaToBaseMessageChunk(delta: Record<string, any>, rawResponse: OpenAI$1.Chat.Completions.ChatCompletionChunk, defaultRole?: OpenAI$1.Chat.ChatCompletionRole): BaseMessageChunk; /** * @deprecated * This function was hoisted into a publicly accessible function from a * different export, but to maintain backwards compatibility with chat models * that depend on ChatOpenAICompletions, we'll keep it here as an overridable * method. This will be removed in a future release */ protected _convertCompletionsMessageToBaseMessage(message: OpenAI$1.ChatCompletionMessage, rawResponse: OpenAI$1.ChatCompletion): BaseMessage; } //#endregion export { ChatOpenAICompletions, ChatOpenAICompletionsCallOptions }; //# sourceMappingURL=completions.d.cts.map