UNPKG

@aws-sdk/client-s3

Version:

AWS SDK for JavaScript S3 Client for Node.js, Browser and React Native

308 lines (307 loc) 19 kB
import { Command as $Command } from "@smithy/smithy-client"; import type { MetadataBearer as __MetadataBearer, StreamingBlobPayloadInputTypes } from "@smithy/types"; import type { UploadPartOutput, UploadPartRequest } from "../models/models_1"; import type { S3ClientResolvedConfig, ServiceInputTypes, ServiceOutputTypes } from "../S3Client"; /** * @public */ export type { __MetadataBearer }; export { $Command }; /** * @public * * The input for {@link UploadPartCommand}. */ export interface UploadPartCommandInput extends Omit<UploadPartRequest, "Body"> { Body?: StreamingBlobPayloadInputTypes; } /** * @public * * The output of {@link UploadPartCommand}. */ export interface UploadPartCommandOutput extends UploadPartOutput, __MetadataBearer { } declare const UploadPartCommand_base: { new (input: UploadPartCommandInput): import("@smithy/smithy-client").CommandImpl<UploadPartCommandInput, UploadPartCommandOutput, S3ClientResolvedConfig, ServiceInputTypes, ServiceOutputTypes>; new (input: UploadPartCommandInput): import("@smithy/smithy-client").CommandImpl<UploadPartCommandInput, UploadPartCommandOutput, S3ClientResolvedConfig, ServiceInputTypes, ServiceOutputTypes>; getEndpointParameterInstructions(): import("@smithy/middleware-endpoint").EndpointParameterInstructions; }; /** * <p>Uploads a part in a multipart upload.</p> * <note> * <p>In this operation, you provide new data as a part of an object in your request. However, you have * an option to specify your existing Amazon S3 object as a data source for the part you are uploading. To * upload a part from an existing object, you use the <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_UploadPartCopy.html">UploadPartCopy</a> operation. </p> * </note> * <p>You must initiate a multipart upload (see <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CreateMultipartUpload.html">CreateMultipartUpload</a>) before you can * upload any part. In response to your initiate request, Amazon S3 returns an upload ID, a unique identifier * that you must include in your upload part request.</p> * <p>Part numbers can be any number from 1 to 10,000, inclusive. A part number uniquely identifies a part * and also defines its position within the object being created. If you upload a new part using the same * part number that was used with a previous part, the previously uploaded part is overwritten.</p> * <p>For information about maximum and minimum part sizes and other multipart upload specifications, see * <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/qfacts.html">Multipart upload * limits</a> in the <i>Amazon S3 User Guide</i>.</p> * <note> * <p>After you initiate multipart upload and upload one or more parts, you must either complete or * abort multipart upload in order to stop getting charged for storage of the uploaded parts. Only after * you either complete or abort multipart upload, Amazon S3 frees up the parts storage and stops charging you * for the parts storage.</p> * </note> * <p>For more information on multipart uploads, go to <a href="https://docs.aws.amazon.com/AmazonS3/latest/dev/mpuoverview.html">Multipart Upload Overview</a> in the * <i>Amazon S3 User Guide </i>.</p> * <note> * <p> * <b>Directory buckets</b> - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format <code>https://<i>amzn-s3-demo-bucket</i>.s3express-<i>zone-id</i>.<i>region-code</i>.amazonaws.com/<i>key-name</i> * </code>. Path-style requests are not supported. For more information about endpoints in Availability Zones, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/endpoint-directory-buckets-AZ.html">Regional and Zonal endpoints for directory buckets in Availability Zones</a> in the * <i>Amazon S3 User Guide</i>. For more information about endpoints in Local Zones, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/s3-lzs-for-directory-buckets.html">Concepts for directory buckets in Local Zones</a> in the * <i>Amazon S3 User Guide</i>.</p> * </note> * <dl> * <dt>Permissions</dt> * <dd> * <ul> * <li> * <p> * <b>General purpose bucket permissions</b> - To perform a * multipart upload with encryption using an Key Management Service key, the requester must have permission to * the <code>kms:Decrypt</code> and <code>kms:GenerateDataKey</code> actions on the key. The * requester must also have permissions for the <code>kms:GenerateDataKey</code> action for the * <code>CreateMultipartUpload</code> API. Then, the requester needs permissions for the * <code>kms:Decrypt</code> action on the <code>UploadPart</code> and * <code>UploadPartCopy</code> APIs.</p> * <p>These permissions are required because Amazon S3 must decrypt and read data from the encrypted * file parts before it completes the multipart upload. For more information about KMS * permissions, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/UsingKMSEncryption.html">Protecting data using server-side * encryption with KMS</a> in the <i>Amazon S3 User Guide</i>. For information * about the permissions required to use the multipart upload API, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/dev/mpuAndPermissions.html">Multipart upload and * permissions</a> and <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/mpuoverview.html#mpuAndPermissions">Multipart upload API and * permissions</a> in the <i>Amazon S3 User Guide</i>.</p> * </li> * <li> * <p> * <b>Directory bucket permissions</b> - To grant access to this API operation on a directory bucket, we recommend that you use the <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CreateSession.html"> * <code>CreateSession</code> * </a> API operation for session-based authorization. Specifically, you grant the <code>s3express:CreateSession</code> permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the <code>CreateSession</code> API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another <code>CreateSession</code> API call to generate a new session token for use. * Amazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CreateSession.html"> * <code>CreateSession</code> * </a>.</p> * <p>If the object is encrypted with SSE-KMS, you must also have the * <code>kms:GenerateDataKey</code> and <code>kms:Decrypt</code> permissions in IAM * identity-based policies and KMS key policies for the KMS key.</p> * </li> * </ul> * </dd> * <dt>Data integrity</dt> * <dd> * <p> * <b>General purpose bucket</b> - To ensure that data is not corrupted * traversing the network, specify the <code>Content-MD5</code> header in the upload part request. * Amazon S3 checks the part data against the provided MD5 value. If they do not match, Amazon S3 returns an * error. If the upload request is signed with Signature Version 4, then Amazon Web Services S3 uses the * <code>x-amz-content-sha256</code> header as a checksum instead of <code>Content-MD5</code>. For * more information see <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/sigv4-auth-using-authorization-header.html">Authenticating Requests: * Using the Authorization Header (Amazon Web Services Signature Version 4)</a>. </p> * <note> * <p> * <b>Directory buckets</b> - MD5 is not supported by directory buckets. You can use checksum algorithms to check object integrity.</p> * </note> * </dd> * <dt>Encryption</dt> * <dd> * <ul> * <li> * <p> * <b>General purpose bucket</b> - Server-side encryption is for * data encryption at rest. Amazon S3 encrypts your data as it writes it to disks in its data centers * and decrypts it when you access it. You have mutually exclusive options to protect data using * server-side encryption in Amazon S3, depending on how you choose to manage the encryption keys. * Specifically, the encryption key options are Amazon S3 managed keys (SSE-S3), Amazon Web Services KMS keys * (SSE-KMS), and Customer-Provided Keys (SSE-C). Amazon S3 encrypts data with server-side encryption * using Amazon S3 managed keys (SSE-S3) by default. You can optionally tell Amazon S3 to encrypt data at * rest using server-side encryption with other key options. The option you use depends on * whether you want to use KMS keys (SSE-KMS) or provide your own encryption key * (SSE-C).</p> * <p>Server-side encryption is supported by the S3 Multipart Upload operations. Unless you are * using a customer-provided encryption key (SSE-C), you don't need to specify the encryption * parameters in each UploadPart request. Instead, you only need to specify the server-side * encryption parameters in the initial Initiate Multipart request. For more information, see * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CreateMultipartUpload.html">CreateMultipartUpload</a>.</p> * <note> * <p>If you have server-side encryption with customer-provided keys (SSE-C) blocked for your general purpose bucket, you will get an HTTP 403 Access Denied error when you specify the SSE-C request headers while writing new data to your bucket. For more information, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/userguide/blocking-unblocking-s3-c-encryption-gpb.html">Blocking or unblocking SSE-C for a general purpose bucket</a>.</p> * </note> * <p>If you request server-side encryption using a customer-provided encryption key (SSE-C) in * your initiate multipart upload request, you must provide identical encryption information in * each part upload using the following request headers.</p> * <ul> * <li> * <p>x-amz-server-side-encryption-customer-algorithm</p> * </li> * <li> * <p>x-amz-server-side-encryption-customer-key</p> * </li> * <li> * <p>x-amz-server-side-encryption-customer-key-MD5</p> * </li> * </ul> * <p> For more information, see <a href="https://docs.aws.amazon.com/AmazonS3/latest/dev/UsingServerSideEncryption.html">Using Server-Side * Encryption</a> in the <i>Amazon S3 User Guide</i>.</p> * </li> * <li> * <p> * <b>Directory buckets </b> - * For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (<code>AES256</code>) and server-side encryption with KMS keys (SSE-KMS) (<code>aws:kms</code>).</p> * </li> * </ul> * </dd> * <dt>Special errors</dt> * <dd> * <ul> * <li> * <p>Error Code: <code>NoSuchUpload</code> * </p> * <ul> * <li> * <p>Description: The specified multipart upload does not exist. The upload ID might be * invalid, or the multipart upload might have been aborted or completed.</p> * </li> * <li> * <p>HTTP Status Code: 404 Not Found </p> * </li> * <li> * <p>SOAP Fault Code Prefix: Client</p> * </li> * </ul> * </li> * </ul> * </dd> * <dt>HTTP Host header syntax</dt> * <dd> * <p> * <b>Directory buckets </b> - The HTTP Host header syntax is <code> * <i>Bucket-name</i>.s3express-<i>zone-id</i>.<i>region-code</i>.amazonaws.com</code>.</p> * </dd> * </dl> * <p>The following operations are related to <code>UploadPart</code>:</p> * <ul> * <li> * <p> * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CreateMultipartUpload.html">CreateMultipartUpload</a> * </p> * </li> * <li> * <p> * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_CompleteMultipartUpload.html">CompleteMultipartUpload</a> * </p> * </li> * <li> * <p> * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_AbortMultipartUpload.html">AbortMultipartUpload</a> * </p> * </li> * <li> * <p> * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_ListParts.html">ListParts</a> * </p> * </li> * <li> * <p> * <a href="https://docs.aws.amazon.com/AmazonS3/latest/API/API_ListMultipartUploads.html">ListMultipartUploads</a> * </p> * </li> * </ul> * <important> * <p>You must URL encode any signed header values that contain spaces. For example, if your header value is <code>my file.txt</code>, containing two spaces after <code>my</code>, you must URL encode this value to <code>my%20%20file.txt</code>.</p> * </important> * @example * Use a bare-bones client and the command you need to make an API call. * ```javascript * import { S3Client, UploadPartCommand } from "@aws-sdk/client-s3"; // ES Modules import * // const { S3Client, UploadPartCommand } = require("@aws-sdk/client-s3"); // CommonJS import * // import type { S3ClientConfig } from "@aws-sdk/client-s3"; * const config = {}; // type is S3ClientConfig * const client = new S3Client(config); * const input = { // UploadPartRequest * Body: "MULTIPLE_TYPES_ACCEPTED", // see \@smithy/types -> StreamingBlobPayloadInputTypes * Bucket: "STRING_VALUE", // required * ContentLength: Number("long"), * ContentMD5: "STRING_VALUE", * ChecksumAlgorithm: "CRC32" || "CRC32C" || "SHA1" || "SHA256" || "CRC64NVME", * ChecksumCRC32: "STRING_VALUE", * ChecksumCRC32C: "STRING_VALUE", * ChecksumCRC64NVME: "STRING_VALUE", * ChecksumSHA1: "STRING_VALUE", * ChecksumSHA256: "STRING_VALUE", * Key: "STRING_VALUE", // required * PartNumber: Number("int"), // required * UploadId: "STRING_VALUE", // required * SSECustomerAlgorithm: "STRING_VALUE", * SSECustomerKey: "STRING_VALUE", * SSECustomerKeyMD5: "STRING_VALUE", * RequestPayer: "requester", * ExpectedBucketOwner: "STRING_VALUE", * }; * const command = new UploadPartCommand(input); * const response = await client.send(command); * // { // UploadPartOutput * // ServerSideEncryption: "AES256" || "aws:fsx" || "aws:kms" || "aws:kms:dsse", * // ETag: "STRING_VALUE", * // ChecksumCRC32: "STRING_VALUE", * // ChecksumCRC32C: "STRING_VALUE", * // ChecksumCRC64NVME: "STRING_VALUE", * // ChecksumSHA1: "STRING_VALUE", * // ChecksumSHA256: "STRING_VALUE", * // SSECustomerAlgorithm: "STRING_VALUE", * // SSECustomerKeyMD5: "STRING_VALUE", * // SSEKMSKeyId: "STRING_VALUE", * // BucketKeyEnabled: true || false, * // RequestCharged: "requester", * // }; * * ``` * * @param UploadPartCommandInput - {@link UploadPartCommandInput} * @returns {@link UploadPartCommandOutput} * @see {@link UploadPartCommandInput} for command's `input` shape. * @see {@link UploadPartCommandOutput} for command's `response` shape. * @see {@link S3ClientResolvedConfig | config} for S3Client's `config` shape. * * @throws {@link S3ServiceException} * <p>Base exception class for all service exceptions from S3 service.</p> * * * @example To upload a part * ```javascript * // The following example uploads part 1 of a multipart upload. The example specifies a file name for the part data. The Upload ID is same that is returned by the initiate multipart upload. * const input = { * Body: "fileToUpload", * Bucket: "examplebucket", * Key: "examplelargeobject", * PartNumber: 1, * UploadId: "xadcOB_7YPBOJuoFiQ9cz4P3Pe6FIZwO4f7wN93uHsNBEw97pl5eNwzExg0LAT2dUN91cOmrEQHDsP3WA60CEg--" * }; * const command = new UploadPartCommand(input); * const response = await client.send(command); * /* response is * { * ETag: `"d8c2eafd90c266e19ab9dcacc479f8af"` * } * *\/ * ``` * * @public */ export declare class UploadPartCommand extends UploadPartCommand_base { /** @internal type navigation helper, not in runtime. */ protected static __types: { api: { input: UploadPartRequest; output: UploadPartOutput; }; sdk: { input: UploadPartCommandInput; output: UploadPartCommandOutput; }; }; }