UNPKG

@anthropic-ai/sdk

Version:

The official TypeScript library for the Anthropic API

71 lines 3.29 kB
"use strict"; // File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. Object.defineProperty(exports, "__esModule", { value: true }); exports.MessageBatchesPage = exports.Batches = void 0; const resource_1 = require("../../resource.js"); const core_1 = require("../../core.js"); const pagination_1 = require("../../pagination.js"); const jsonl_1 = require("../../internal/decoders/jsonl.js"); const error_1 = require("../../error.js"); class Batches extends resource_1.APIResource { /** * Send a batch of Message creation requests. * * The Message Batches API can be used to process multiple Messages API requests at * once. Once a Message Batch is created, it begins processing immediately. Batches * can take up to 24 hours to complete. */ create(body, options) { return this._client.post('/v1/messages/batches', { body, ...options }); } /** * This endpoint is idempotent and can be used to poll for Message Batch * completion. To access the results of a Message Batch, make a request to the * `results_url` field in the response. */ retrieve(messageBatchId, options) { return this._client.get(`/v1/messages/batches/${messageBatchId}`, options); } list(query = {}, options) { if ((0, core_1.isRequestOptions)(query)) { return this.list({}, query); } return this._client.getAPIList('/v1/messages/batches', MessageBatchesPage, { query, ...options }); } /** * Batches may be canceled any time before processing ends. Once cancellation is * initiated, the batch enters a `canceling` state, at which time the system may * complete any in-progress, non-interruptible requests before finalizing * cancellation. * * The number of canceled requests is specified in `request_counts`. To determine * which requests were canceled, check the individual results within the batch. * Note that cancellation may not result in any canceled requests if they were * non-interruptible. */ cancel(messageBatchId, options) { return this._client.post(`/v1/messages/batches/${messageBatchId}/cancel`, options); } /** * Streams the results of a Message Batch as a `.jsonl` file. * * Each line in the file is a JSON object containing the result of a single request * in the Message Batch. Results are not guaranteed to be in the same order as * requests. Use the `custom_id` field to match results to requests. */ async results(messageBatchId, options) { const batch = await this.retrieve(messageBatchId); if (!batch.results_url) { throw new error_1.AnthropicError(`No batch \`results_url\`; Has it finished processing? ${batch.processing_status} - ${batch.id}`); } return this._client .get(batch.results_url, { ...options, __binaryResponse: true }) ._thenUnwrap((_, props) => jsonl_1.JSONLDecoder.fromResponse(props.response, props.controller)); } } exports.Batches = Batches; class MessageBatchesPage extends pagination_1.Page { } exports.MessageBatchesPage = MessageBatchesPage; Batches.MessageBatchesPage = MessageBatchesPage; //# sourceMappingURL=batches.js.map