UNPKG

openai

Version:

The official TypeScript library for the OpenAI API

503 lines (434 loc) 19.8 kB
// File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. import * as Core from './core'; import * as Errors from './error'; import { type Agent, type RequestInit } from './_shims/index'; import * as Uploads from './uploads'; import * as Pagination from "./pagination"; import * as API from "./resources/index"; export interface ClientOptions { /** * Defaults to process.env['OPENAI_API_KEY']. */ apiKey?: string | undefined; /** * Defaults to process.env['OPENAI_ORG_ID']. */ organization?: string | null | undefined; /** * Defaults to process.env['OPENAI_PROJECT_ID']. */ project?: string | null | undefined; /** * Override the default base URL for the API, e.g., "https://api.example.com/v2/" * * Defaults to process.env['OPENAI_BASE_URL']. */ baseURL?: string | null | undefined; /** * The maximum amount of time (in milliseconds) that the client should wait for a response * from the server before timing out a single request. * * Note that request timeouts are retried by default, so in a worst-case scenario you may wait * much longer than this timeout before the promise succeeds or fails. */ timeout?: number; /** * An HTTP agent used to manage HTTP(S) connections. * * If not provided, an agent will be constructed by default in the Node.js environment, * otherwise no agent is used. */ httpAgent?: Agent; /** * Specify a custom `fetch` function implementation. * * If not provided, we use `node-fetch` on Node.js and otherwise expect that `fetch` is * defined globally. */ fetch?: Core.Fetch | undefined; /** * The maximum number of times that the client will retry a request in case of a * temporary failure, like a network error or a 5XX error from the server. * * @default 2 */ maxRetries?: number; /** * Default headers to include with every request to the API. * * These can be removed in individual requests by explicitly setting the * header to `undefined` or `null` in request options. */ defaultHeaders?: Core.Headers; /** * Default query parameters to include with every request to the API. * * These can be removed in individual requests by explicitly setting the * param to `undefined` in request options. */ defaultQuery?: Core.DefaultQuery; /** * By default, client-side use of this library is not allowed, as it risks exposing your secret API credentials to attackers. * Only set this option to `true` if you understand the risks and have appropriate mitigations in place. */ dangerouslyAllowBrowser?: boolean; } /** API Client for interfacing with the OpenAI API. */ export class OpenAI extends Core.APIClient { apiKey: string; organization: string | null; project: string | null; private _options: ClientOptions; /** * API Client for interfacing with the OpenAI API. * * @param {string | undefined} [opts.apiKey=process.env['OPENAI_API_KEY'] ?? undefined] * @param {string | null | undefined} [opts.organization=process.env['OPENAI_ORG_ID'] ?? null] * @param {string | null | undefined} [opts.project=process.env['OPENAI_PROJECT_ID'] ?? null] * @param {string} [opts.baseURL=process.env['OPENAI_BASE_URL'] ?? https://api.openai.com/v1] - Override the default base URL for the API. * @param {number} [opts.timeout=10 minutes] - The maximum amount of time (in milliseconds) the client will wait for a response before timing out. * @param {number} [opts.httpAgent] - An HTTP agent used to manage HTTP(s) connections. * @param {Core.Fetch} [opts.fetch] - Specify a custom `fetch` function implementation. * @param {number} [opts.maxRetries=2] - The maximum number of times the client will retry a request. * @param {Core.Headers} opts.defaultHeaders - Default headers to include with every request to the API. * @param {Core.DefaultQuery} opts.defaultQuery - Default query parameters to include with every request to the API. * @param {boolean} [opts.dangerouslyAllowBrowser=false] - By default, client-side use of this library is not allowed, as it risks exposing your secret API credentials to attackers. */ constructor({ baseURL = Core.readEnv('OPENAI_BASE_URL'), apiKey = Core.readEnv('OPENAI_API_KEY'), organization = Core.readEnv('OPENAI_ORG_ID') ?? null, project = Core.readEnv('OPENAI_PROJECT_ID') ?? null, ...opts }: ClientOptions = {}) { if (apiKey === undefined) { throw new Errors.OpenAIError( "The OPENAI_API_KEY environment variable is missing or empty; either provide it, or instantiate the OpenAI client with an apiKey option, like new OpenAI({ apiKey: 'My API Key' }).", ); } const options: ClientOptions = { apiKey, organization, project, ...opts, baseURL: baseURL || `https://api.openai.com/v1`, }; if (!options.dangerouslyAllowBrowser && Core.isRunningInBrowser()) { throw new Errors.OpenAIError( "It looks like you're running in a browser-like environment.\n\nThis is disabled by default, as it risks exposing your secret API credentials to attackers.\nIf you understand the risks and have appropriate mitigations in place,\nyou can set the `dangerouslyAllowBrowser` option to `true`, e.g.,\n\nnew OpenAI({ apiKey, dangerouslyAllowBrowser: true });\n\nhttps://help.openai.com/en/articles/5112595-best-practices-for-api-key-safety\n", ); } super({ baseURL: options.baseURL!, timeout: options.timeout ?? 600000 /* 10 minutes */, httpAgent: options.httpAgent, maxRetries: options.maxRetries, fetch: options.fetch, }); this._options = options; this.apiKey = apiKey; this.organization = organization; this.project = project; } completions: API.Completions = new API.Completions(this); chat: API.Chat = new API.Chat(this); embeddings: API.Embeddings = new API.Embeddings(this); files: API.Files = new API.Files(this); images: API.Images = new API.Images(this); audio: API.Audio = new API.Audio(this); moderations: API.Moderations = new API.Moderations(this); models: API.Models = new API.Models(this); fineTuning: API.FineTuning = new API.FineTuning(this); beta: API.Beta = new API.Beta(this); batches: API.Batches = new API.Batches(this); protected override defaultQuery(): Core.DefaultQuery | undefined { return this._options.defaultQuery; } protected override defaultHeaders(opts: Core.FinalRequestOptions): Core.Headers { return { ...super.defaultHeaders(opts), 'OpenAI-Organization': this.organization, 'OpenAI-Project': this.project, ...this._options.defaultHeaders, }; } protected override authHeaders(opts: Core.FinalRequestOptions): Core.Headers { return { Authorization: `Bearer ${this.apiKey}` }; } static OpenAI = this; static OpenAIError = Errors.OpenAIError; static APIError = Errors.APIError; static APIConnectionError = Errors.APIConnectionError; static APIConnectionTimeoutError = Errors.APIConnectionTimeoutError; static APIUserAbortError = Errors.APIUserAbortError; static NotFoundError = Errors.NotFoundError; static ConflictError = Errors.ConflictError; static RateLimitError = Errors.RateLimitError; static BadRequestError = Errors.BadRequestError; static AuthenticationError = Errors.AuthenticationError; static InternalServerError = Errors.InternalServerError; static PermissionDeniedError = Errors.PermissionDeniedError; static UnprocessableEntityError = Errors.UnprocessableEntityError; static toFile = Uploads.toFile; static fileFromPath = Uploads.fileFromPath; } export const { OpenAIError, APIError, APIConnectionError, APIConnectionTimeoutError, APIUserAbortError, NotFoundError, ConflictError, RateLimitError, BadRequestError, AuthenticationError, InternalServerError, PermissionDeniedError, UnprocessableEntityError, } = Errors; export import toFile = Uploads.toFile; export import fileFromPath = Uploads.fileFromPath; export namespace OpenAI { export import RequestOptions = Core.RequestOptions; export import Page = Pagination.Page; export import PageResponse = Pagination.PageResponse; export import CursorPage = Pagination.CursorPage; export import CursorPageParams = Pagination.CursorPageParams; export import CursorPageResponse = Pagination.CursorPageResponse; export import Completions = API.Completions; export import Completion = API.Completion; export import CompletionChoice = API.CompletionChoice; export import CompletionUsage = API.CompletionUsage; export import CompletionCreateParams = API.CompletionCreateParams; export import CompletionCreateParamsNonStreaming = API.CompletionCreateParamsNonStreaming; export import CompletionCreateParamsStreaming = API.CompletionCreateParamsStreaming; export import Chat = API.Chat; export import ChatModel = API.ChatModel; export import ChatCompletion = API.ChatCompletion; export import ChatCompletionAssistantMessageParam = API.ChatCompletionAssistantMessageParam; export import ChatCompletionChunk = API.ChatCompletionChunk; export import ChatCompletionContentPart = API.ChatCompletionContentPart; export import ChatCompletionContentPartImage = API.ChatCompletionContentPartImage; export import ChatCompletionContentPartText = API.ChatCompletionContentPartText; export import ChatCompletionFunctionCallOption = API.ChatCompletionFunctionCallOption; export import ChatCompletionFunctionMessageParam = API.ChatCompletionFunctionMessageParam; export import ChatCompletionMessage = API.ChatCompletionMessage; export import ChatCompletionMessageParam = API.ChatCompletionMessageParam; export import ChatCompletionMessageToolCall = API.ChatCompletionMessageToolCall; export import ChatCompletionNamedToolChoice = API.ChatCompletionNamedToolChoice; export import ChatCompletionRole = API.ChatCompletionRole; export import ChatCompletionStreamOptions = API.ChatCompletionStreamOptions; export import ChatCompletionSystemMessageParam = API.ChatCompletionSystemMessageParam; export import ChatCompletionTokenLogprob = API.ChatCompletionTokenLogprob; export import ChatCompletionTool = API.ChatCompletionTool; export import ChatCompletionToolChoiceOption = API.ChatCompletionToolChoiceOption; export import ChatCompletionToolMessageParam = API.ChatCompletionToolMessageParam; export import ChatCompletionUserMessageParam = API.ChatCompletionUserMessageParam; export import ChatCompletionCreateParams = API.ChatCompletionCreateParams; export import ChatCompletionCreateParamsNonStreaming = API.ChatCompletionCreateParamsNonStreaming; export import ChatCompletionCreateParamsStreaming = API.ChatCompletionCreateParamsStreaming; export import Embeddings = API.Embeddings; export import CreateEmbeddingResponse = API.CreateEmbeddingResponse; export import Embedding = API.Embedding; export import EmbeddingCreateParams = API.EmbeddingCreateParams; export import Files = API.Files; export import FileContent = API.FileContent; export import FileDeleted = API.FileDeleted; export import FileObject = API.FileObject; export import FileObjectsPage = API.FileObjectsPage; export import FileCreateParams = API.FileCreateParams; export import FileListParams = API.FileListParams; export import Images = API.Images; export import Image = API.Image; export import ImagesResponse = API.ImagesResponse; export import ImageCreateVariationParams = API.ImageCreateVariationParams; export import ImageEditParams = API.ImageEditParams; export import ImageGenerateParams = API.ImageGenerateParams; export import Audio = API.Audio; export import Moderations = API.Moderations; export import Moderation = API.Moderation; export import ModerationCreateResponse = API.ModerationCreateResponse; export import ModerationCreateParams = API.ModerationCreateParams; export import Models = API.Models; export import Model = API.Model; export import ModelDeleted = API.ModelDeleted; export import ModelsPage = API.ModelsPage; export import FineTuning = API.FineTuning; export import Beta = API.Beta; export import Batches = API.Batches; export import Batch = API.Batch; export import BatchError = API.BatchError; export import BatchRequestCounts = API.BatchRequestCounts; export import BatchesPage = API.BatchesPage; export import BatchCreateParams = API.BatchCreateParams; export import BatchListParams = API.BatchListParams; export import ErrorObject = API.ErrorObject; export import FunctionDefinition = API.FunctionDefinition; export import FunctionParameters = API.FunctionParameters; } // ---------------------- Azure ---------------------- /** API Client for interfacing with the Azure OpenAI API. */ export interface AzureClientOptions extends ClientOptions { /** * Defaults to process.env['OPENAI_API_VERSION']. */ apiVersion?: string | undefined; /** * Your Azure endpoint, including the resource, e.g. `https://example-resource.azure.openai.com/` */ endpoint?: string | undefined; /** * A model deployment, if given, sets the base client URL to include `/deployments/{deployment}`. * Note: this means you won't be able to use non-deployment endpoints. Not supported with Assistants APIs. */ deployment?: string | undefined; /** * Defaults to process.env['AZURE_OPENAI_API_KEY']. */ apiKey?: string | undefined; /** * A function that returns an access token for Microsoft Entra (formerly known as Azure Active Directory), * which will be invoked on every request. */ azureADTokenProvider?: (() => Promise<string>) | undefined; } /** API Client for interfacing with the Azure OpenAI API. */ export class AzureOpenAI extends OpenAI { private _azureADTokenProvider: (() => Promise<string>) | undefined; private _deployment: string | undefined; apiVersion: string = ''; /** * API Client for interfacing with the Azure OpenAI API. * * @param {string | undefined} [opts.apiVersion=process.env['OPENAI_API_VERSION'] ?? undefined] * @param {string | undefined} [opts.endpoint=process.env['AZURE_OPENAI_ENDPOINT'] ?? undefined] - Your Azure endpoint, including the resource, e.g. `https://example-resource.azure.openai.com/` * @param {string | undefined} [opts.apiKey=process.env['AZURE_OPENAI_API_KEY'] ?? undefined] * @param {string | undefined} opts.deployment - A model deployment, if given, sets the base client URL to include `/deployments/{deployment}`. * @param {string | null | undefined} [opts.organization=process.env['OPENAI_ORG_ID'] ?? null] * @param {string} [opts.baseURL=process.env['OPENAI_BASE_URL']] - Sets the base URL for the API. * @param {number} [opts.timeout=10 minutes] - The maximum amount of time (in milliseconds) the client will wait for a response before timing out. * @param {number} [opts.httpAgent] - An HTTP agent used to manage HTTP(s) connections. * @param {Core.Fetch} [opts.fetch] - Specify a custom `fetch` function implementation. * @param {number} [opts.maxRetries=2] - The maximum number of times the client will retry a request. * @param {Core.Headers} opts.defaultHeaders - Default headers to include with every request to the API. * @param {Core.DefaultQuery} opts.defaultQuery - Default query parameters to include with every request to the API. * @param {boolean} [opts.dangerouslyAllowBrowser=false] - By default, client-side use of this library is not allowed, as it risks exposing your secret API credentials to attackers. */ constructor({ baseURL = Core.readEnv('OPENAI_BASE_URL'), apiKey = Core.readEnv('AZURE_OPENAI_API_KEY'), apiVersion = Core.readEnv('OPENAI_API_VERSION'), endpoint, deployment, azureADTokenProvider, dangerouslyAllowBrowser, ...opts }: AzureClientOptions = {}) { if (!apiVersion) { throw new Errors.OpenAIError( "The OPENAI_API_VERSION environment variable is missing or empty; either provide it, or instantiate the AzureOpenAI client with an apiVersion option, like new AzureOpenAI({ apiVersion: 'My API Version' }).", ); } if (typeof azureADTokenProvider === 'function') { dangerouslyAllowBrowser = true; } if (!azureADTokenProvider && !apiKey) { throw new Errors.OpenAIError( 'Missing credentials. Please pass one of `apiKey` and `azureADTokenProvider`, or set the `AZURE_OPENAI_API_KEY` environment variable.', ); } if (azureADTokenProvider && apiKey) { throw new Errors.OpenAIError( 'The `apiKey` and `azureADTokenProvider` arguments are mutually exclusive; only one can be passed at a time.', ); } // define a sentinel value to avoid any typing issues apiKey ??= API_KEY_SENTINEL; opts.defaultQuery = { ...opts.defaultQuery, 'api-version': apiVersion }; if (!baseURL) { if (!endpoint) { endpoint = process.env['AZURE_OPENAI_ENDPOINT']; } if (!endpoint) { throw new Errors.OpenAIError( 'Must provide one of the `baseURL` or `endpoint` arguments, or the `AZURE_OPENAI_ENDPOINT` environment variable', ); } baseURL = `${endpoint}/openai`; } else { if (endpoint) { throw new Errors.OpenAIError('baseURL and endpoint are mutually exclusive'); } } super({ apiKey, baseURL, ...opts, ...(dangerouslyAllowBrowser !== undefined ? { dangerouslyAllowBrowser } : {}), }); this._azureADTokenProvider = azureADTokenProvider; this.apiVersion = apiVersion; this._deployment = deployment; } override buildRequest(options: Core.FinalRequestOptions<unknown>): { req: RequestInit; url: string; timeout: number; } { if (_deployments_endpoints.has(options.path) && options.method === 'post' && options.body !== undefined) { if (!Core.isObj(options.body)) { throw new Error('Expected request body to be an object'); } const model = this._deployment || options.body['model']; delete options.body['model']; if (model !== undefined && !this.baseURL.includes('/deployments')) { options.path = `/deployments/${model}${options.path}`; } } return super.buildRequest(options); } private async _getAzureADToken(): Promise<string | undefined> { if (typeof this._azureADTokenProvider === 'function') { const token = await this._azureADTokenProvider(); if (!token || typeof token !== 'string') { throw new Errors.OpenAIError( `Expected 'azureADTokenProvider' argument to return a string but it returned ${token}`, ); } return token; } return undefined; } protected override authHeaders(opts: Core.FinalRequestOptions): Core.Headers { return {}; } protected override async prepareOptions(opts: Core.FinalRequestOptions<unknown>): Promise<void> { if (opts.headers?.['Authorization'] || opts.headers?.['api-key']) { return super.prepareOptions(opts); } const token = await this._getAzureADToken(); opts.headers ??= {}; if (token) { opts.headers['Authorization'] = `Bearer ${token}`; } else if (this.apiKey !== API_KEY_SENTINEL) { opts.headers['api-key'] = this.apiKey; } else { throw new Errors.OpenAIError('Unable to handle auth'); } return super.prepareOptions(opts); } } const _deployments_endpoints = new Set([ '/completions', '/chat/completions', '/embeddings', '/audio/transcriptions', '/audio/translations', '/audio/speech', '/images/generations', '/batches', ]); const API_KEY_SENTINEL = '<Missing Key>'; // ---------------------- End Azure ---------------------- export default OpenAI;