From d3b2e8c537c679318dfefbfbbf7ffb1bcdddf655 Mon Sep 17 00:00:00 2001
From: awstools
You specify the lifecycle configuration in your request body. The lifecycle - * configuration is specified as XML consisting of one or more rules. Each rule consists of - * the following:
+ * configuration is specified as XML consisting of one or more rules. An Amazon S3 Lifecycle + * configuration can have up to 1,000 rules. This limit is not adjustable. Each rule consists + * of the following: * *Part numbers can be any number from 1 to 10,000, inclusive. A part number uniquely * identifies a part and also defines its position within the object being created. If you * upload a new part using the same part number that was used with a previous part, the - * previously uploaded part is overwritten. Each part must be at least 5 MB in size, except - * the last part. There is no size limit on the last part of your multipart upload.
+ * previously uploaded part is overwritten. + *For information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
*To ensure that data is not corrupted when traversing the network, specify the
* Content-MD5
header in the upload part request. Amazon S3 checks the part data
* against the provided MD5 value. If they do not match, Amazon S3 returns an error.
x-amz-copy-source
in your request and
* a byte range by adding the request header x-amz-copy-source-range
in your
* request.
- * The minimum allowable part size for a multipart upload is 5 MB. For more information - * about multipart upload limits, go to Quick - * Facts in the Amazon S3 User Guide.
+ *For information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
*Instead of using an existing object as part data, you might use the UploadPart * action and provide data in your request.
diff --git a/clients/client-s3/src/commands/PutBucketLifecycleConfigurationCommand.ts b/clients/client-s3/src/commands/PutBucketLifecycleConfigurationCommand.ts index 1a1b1b97d218..3ef295c358a1 100644 --- a/clients/client-s3/src/commands/PutBucketLifecycleConfigurationCommand.ts +++ b/clients/client-s3/src/commands/PutBucketLifecycleConfigurationCommand.ts @@ -44,8 +44,9 @@ export interface PutBucketLifecycleConfigurationCommandOutput extends __Metadata * Rules * *You specify the lifecycle configuration in your request body. The lifecycle - * configuration is specified as XML consisting of one or more rules. Each rule consists of - * the following:
+ * configuration is specified as XML consisting of one or more rules. An Amazon S3 Lifecycle + * configuration can have up to 1,000 rules. This limit is not adjustable. Each rule consists + * of the following: * *Part numbers can be any number from 1 to 10,000, inclusive. A part number uniquely * identifies a part and also defines its position within the object being created. If you * upload a new part using the same part number that was used with a previous part, the - * previously uploaded part is overwritten. Each part must be at least 5 MB in size, except - * the last part. There is no size limit on the last part of your multipart upload.
+ * previously uploaded part is overwritten. + *For information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
*To ensure that data is not corrupted when traversing the network, specify the
* Content-MD5
header in the upload part request. Amazon S3 checks the part data
* against the provided MD5 value. If they do not match, Amazon S3 returns an error.
x-amz-copy-source
in your request and
* a byte range by adding the request header x-amz-copy-source-range
in your
* request.
- * The minimum allowable part size for a multipart upload is 5 MB. For more information - * about multipart upload limits, go to Quick - * Facts in the Amazon S3 User Guide.
+ *For information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
*Instead of using an existing object as part data, you might use the UploadPart * action and provide data in your request.
diff --git a/clients/client-s3/src/models/models_0.ts b/clients/client-s3/src/models/models_0.ts index 43c259bffdc2..91513250665c 100644 --- a/clients/client-s3/src/models/models_0.ts +++ b/clients/client-s3/src/models/models_0.ts @@ -5439,7 +5439,7 @@ export namespace LifecycleRuleFilter { export interface NoncurrentVersionExpiration { /** *Specifies the number of days an object is noncurrent before Amazon S3 can perform the - * associated action. For information about the noncurrent days calculations, see How + * associated action. The value must be a non-zero positive integer. For information about the noncurrent days calculations, see How * Amazon S3 Calculates When an Object Became Noncurrent in the Amazon S3 User Guide.
*/ NoncurrentDays?: number; @@ -9911,7 +9911,9 @@ export interface _Object { *If an object is created by either the Multipart Upload or Part Copy operation, the - * ETag is not an MD5 digest, regardless of the method of encryption.
+ * ETag is not an MD5 digest, regardless of the method of encryption. If an object + * is larger than 16 MB, the Amazon Web Services Management Console will upload or copy that object as a + * Multipart Upload, and therefore the ETag will not be an MD5 digest. *Specifies the number of days an object is noncurrent before Amazon S3 can perform the\n associated action. For information about the noncurrent days calculations, see How\n Amazon S3 Calculates When an Object Became Noncurrent in the Amazon S3 User Guide.
" + "smithy.api#documentation": "Specifies the number of days an object is noncurrent before Amazon S3 can perform the\n associated action. The value must be a non-zero positive integer. For information about the noncurrent days calculations, see How\n Amazon S3 Calculates When an Object Became Noncurrent in the Amazon S3 User Guide.
" } }, "NewerNoncurrentVersions": { @@ -9207,7 +9207,7 @@ "ETag": { "target": "com.amazonaws.s3#ETag", "traits": { - "smithy.api#documentation": "The entity tag is a hash of the object. The ETag reflects changes only to the contents\n of an object, not its metadata. The ETag may or may not be an MD5 digest of the object\n data. Whether or not it is depends on how the object was created and how it is encrypted as\n described below:
\nObjects created by the PUT Object, POST Object, or Copy operation, or through the\n Amazon Web Services Management Console, and are encrypted by SSE-S3 or plaintext, have ETags that are\n an MD5 digest of their object data.
\nObjects created by the PUT Object, POST Object, or Copy operation, or through the\n Amazon Web Services Management Console, and are encrypted by SSE-C or SSE-KMS, have ETags that are\n not an MD5 digest of their object data.
\nIf an object is created by either the Multipart Upload or Part Copy operation, the\n ETag is not an MD5 digest, regardless of the method of encryption.
\nThe entity tag is a hash of the object. The ETag reflects changes only to the contents\n of an object, not its metadata. The ETag may or may not be an MD5 digest of the object\n data. Whether or not it is depends on how the object was created and how it is encrypted as\n described below:
\nObjects created by the PUT Object, POST Object, or Copy operation, or through the\n Amazon Web Services Management Console, and are encrypted by SSE-S3 or plaintext, have ETags that are\n an MD5 digest of their object data.
\nObjects created by the PUT Object, POST Object, or Copy operation, or through the\n Amazon Web Services Management Console, and are encrypted by SSE-C or SSE-KMS, have ETags that are\n not an MD5 digest of their object data.
\nIf an object is created by either the Multipart Upload or Part Copy operation, the\n ETag is not an MD5 digest, regardless of the method of encryption. If an object\n is larger than 16 MB, the Amazon Web Services Management Console will upload or copy that object as a\n Multipart Upload, and therefore the ETag will not be an MD5 digest.
\nCreates a new lifecycle configuration for the bucket or replaces an existing lifecycle\n configuration. Keep in mind that this will overwrite an existing lifecycle configuration, so if\n you want to retain any configuration details, they must be included in the new lifecycle\n configuration. For information about lifecycle configuration, see Managing your storage\n lifecycle.
\n\nBucket lifecycle configuration now supports specifying a lifecycle rule using an\n object key name prefix, one or more object tags, or a combination of both. Accordingly,\n this section describes the latest API. The previous version of the API supported\n filtering based only on an object key name prefix, which is supported for backward\n compatibility. For the related API description, see PutBucketLifecycle.
\n\n Rules\n
\nYou specify the lifecycle configuration in your request body. The lifecycle\n configuration is specified as XML consisting of one or more rules. Each rule consists of\n the following:
\n\nFilter identifying a subset of objects to which the rule applies. The filter can\n be based on a key name prefix, object tags, or a combination of both.
\nStatus whether the rule is in effect.
\nOne or more lifecycle transition and expiration actions that you want Amazon S3 to\n perform on the objects identified by the filter. If the state of your bucket is\n versioning-enabled or versioning-suspended, you can have many versions of the same\n object (one current version and zero or more noncurrent versions). Amazon S3 provides\n predefined actions that you can specify for current and noncurrent object\n versions.
\nFor more information, see Object\n Lifecycle Management and Lifecycle Configuration Elements.
\n\n\n\n Permissions\n
\n\n\nBy default, all Amazon S3 resources are private, including buckets, objects, and related\n subresources (for example, lifecycle configuration and website configuration). Only the\n resource owner (that is, the Amazon Web Services account that created it) can access the resource. The\n resource owner can optionally grant access permissions to others by writing an access\n policy. For this operation, a user must get the s3:PutLifecycleConfiguration
\n permission.
You can also explicitly deny permissions. Explicit deny also supersedes any other\n permissions. If you want to block users or accounts from removing or deleting objects from\n your bucket, you must deny them permissions for the following actions:
\n\n\n s3:DeleteObject
\n
\n s3:DeleteObjectVersion
\n
\n s3:PutLifecycleConfiguration
\n
For more information about permissions, see Managing Access Permissions to Your Amazon S3\n Resources.
\n\nThe following are related to PutBucketLifecycleConfiguration
:
Creates a new lifecycle configuration for the bucket or replaces an existing lifecycle\n configuration. Keep in mind that this will overwrite an existing lifecycle configuration, so if\n you want to retain any configuration details, they must be included in the new lifecycle\n configuration. For information about lifecycle configuration, see Managing your storage\n lifecycle.
\n\nBucket lifecycle configuration now supports specifying a lifecycle rule using an\n object key name prefix, one or more object tags, or a combination of both. Accordingly,\n this section describes the latest API. The previous version of the API supported\n filtering based only on an object key name prefix, which is supported for backward\n compatibility. For the related API description, see PutBucketLifecycle.
\n\n Rules\n
\nYou specify the lifecycle configuration in your request body. The lifecycle\n configuration is specified as XML consisting of one or more rules. An Amazon S3 Lifecycle\n configuration can have up to 1,000 rules. This limit is not adjustable. Each rule consists\n of the following:
\n\nFilter identifying a subset of objects to which the rule applies. The filter can\n be based on a key name prefix, object tags, or a combination of both.
\nStatus whether the rule is in effect.
\nOne or more lifecycle transition and expiration actions that you want Amazon S3 to\n perform on the objects identified by the filter. If the state of your bucket is\n versioning-enabled or versioning-suspended, you can have many versions of the same\n object (one current version and zero or more noncurrent versions). Amazon S3 provides\n predefined actions that you can specify for current and noncurrent object\n versions.
\nFor more information, see Object\n Lifecycle Management and Lifecycle Configuration Elements.
\n\n\n\n Permissions\n
\n\n\nBy default, all Amazon S3 resources are private, including buckets, objects, and related\n subresources (for example, lifecycle configuration and website configuration). Only the\n resource owner (that is, the Amazon Web Services account that created it) can access the resource. The\n resource owner can optionally grant access permissions to others by writing an access\n policy. For this operation, a user must get the s3:PutLifecycleConfiguration
\n permission.
You can also explicitly deny permissions. Explicit deny also supersedes any other\n permissions. If you want to block users or accounts from removing or deleting objects from\n your bucket, you must deny them permissions for the following actions:
\n\n\n s3:DeleteObject
\n
\n s3:DeleteObjectVersion
\n
\n s3:PutLifecycleConfiguration
\n
For more information about permissions, see Managing Access Permissions to Your Amazon S3\n Resources.
\n\nThe following are related to PutBucketLifecycleConfiguration
:
Uploads a part in a multipart upload.
\nIn this operation, you provide part data in your request. However, you have an option\n to specify your existing Amazon S3 object as a data source for the part you are uploading. To\n upload a part from an existing object, you use the UploadPartCopy operation.\n
\nYou must initiate a multipart upload (see CreateMultipartUpload)\n before you can upload any part. In response to your initiate request, Amazon S3 returns an\n upload ID, a unique identifier, that you must include in your upload part request.
\nPart numbers can be any number from 1 to 10,000, inclusive. A part number uniquely\n identifies a part and also defines its position within the object being created. If you\n upload a new part using the same part number that was used with a previous part, the\n previously uploaded part is overwritten. Each part must be at least 5 MB in size, except\n the last part. There is no size limit on the last part of your multipart upload.
\nTo ensure that data is not corrupted when traversing the network, specify the\n Content-MD5
header in the upload part request. Amazon S3 checks the part data\n against the provided MD5 value. If they do not match, Amazon S3 returns an error.
If the upload request is signed with Signature Version 4, then Amazon Web Services S3 uses the\n x-amz-content-sha256
header as a checksum instead of\n Content-MD5
. For more information see Authenticating Requests: Using the Authorization Header (Amazon Web Services Signature Version\n 4).
\n Note: After you initiate multipart upload and upload\n one or more parts, you must either complete or abort multipart upload in order to stop\n getting charged for storage of the uploaded parts. Only after you either complete or abort\n multipart upload, Amazon S3 frees up the parts storage and stops charging you for the parts\n storage.
\n\nFor more information on multipart uploads, go to Multipart Upload Overview in the\n Amazon S3 User Guide .
\nFor information on the permissions required to use the multipart upload API, go to\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\n\nYou can optionally request server-side encryption where Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it for you when you access it. You have\n the option of providing your own encryption key, or you can use the Amazon Web Services managed encryption\n keys. If you choose to provide your own encryption key, the request headers you provide in\n the request must match the headers you used in the request to initiate the upload by using\n CreateMultipartUpload. For more information, go to Using Server-Side Encryption in\n the Amazon S3 User Guide.
\n\nServer-side encryption is supported by the S3 Multipart Upload actions. Unless you are\n using a customer-provided encryption key, you don't need to specify the encryption\n parameters in each UploadPart request. Instead, you only need to specify the server-side\n encryption parameters in the initial Initiate Multipart request. For more information, see\n CreateMultipartUpload.
\n\nIf you requested server-side encryption using a customer-provided encryption key in your\n initiate multipart upload request, you must provide identical encryption information in\n each part upload using the following headers.
\n\n\nx-amz-server-side-encryption-customer-algorithm
\nx-amz-server-side-encryption-customer-key
\nx-amz-server-side-encryption-customer-key-MD5
\n\n Special Errors\n
\n\n Code: NoSuchUpload\n
\n\n Cause: The specified multipart upload does not exist. The upload\n ID might be invalid, or the multipart upload might have been aborted or\n completed.\n
\n\n HTTP Status Code: 404 Not Found \n
\n\n SOAP Fault Code Prefix: Client\n
\n\n Related Resources\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part in a multipart upload.
\nIn this operation, you provide part data in your request. However, you have an option\n to specify your existing Amazon S3 object as a data source for the part you are uploading. To\n upload a part from an existing object, you use the UploadPartCopy operation.\n
\nYou must initiate a multipart upload (see CreateMultipartUpload)\n before you can upload any part. In response to your initiate request, Amazon S3 returns an\n upload ID, a unique identifier, that you must include in your upload part request.
\nPart numbers can be any number from 1 to 10,000, inclusive. A part number uniquely\n identifies a part and also defines its position within the object being created. If you\n upload a new part using the same part number that was used with a previous part, the\n previously uploaded part is overwritten.
\nFor information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nTo ensure that data is not corrupted when traversing the network, specify the\n Content-MD5
header in the upload part request. Amazon S3 checks the part data\n against the provided MD5 value. If they do not match, Amazon S3 returns an error.
If the upload request is signed with Signature Version 4, then Amazon Web Services S3 uses the\n x-amz-content-sha256
header as a checksum instead of\n Content-MD5
. For more information see Authenticating Requests: Using the Authorization Header (Amazon Web Services Signature Version\n 4).
\n Note: After you initiate multipart upload and upload\n one or more parts, you must either complete or abort multipart upload in order to stop\n getting charged for storage of the uploaded parts. Only after you either complete or abort\n multipart upload, Amazon S3 frees up the parts storage and stops charging you for the parts\n storage.
\n\nFor more information on multipart uploads, go to Multipart Upload Overview in the\n Amazon S3 User Guide .
\nFor information on the permissions required to use the multipart upload API, go to\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\n\nYou can optionally request server-side encryption where Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it for you when you access it. You have\n the option of providing your own encryption key, or you can use the Amazon Web Services managed encryption\n keys. If you choose to provide your own encryption key, the request headers you provide in\n the request must match the headers you used in the request to initiate the upload by using\n CreateMultipartUpload. For more information, go to Using Server-Side Encryption in\n the Amazon S3 User Guide.
\n\nServer-side encryption is supported by the S3 Multipart Upload actions. Unless you are\n using a customer-provided encryption key, you don't need to specify the encryption\n parameters in each UploadPart request. Instead, you only need to specify the server-side\n encryption parameters in the initial Initiate Multipart request. For more information, see\n CreateMultipartUpload.
\n\nIf you requested server-side encryption using a customer-provided encryption key in your\n initiate multipart upload request, you must provide identical encryption information in\n each part upload using the following headers.
\n\n\nx-amz-server-side-encryption-customer-algorithm
\nx-amz-server-side-encryption-customer-key
\nx-amz-server-side-encryption-customer-key-MD5
\n\n Special Errors\n
\n\n Code: NoSuchUpload\n
\n\n Cause: The specified multipart upload does not exist. The upload\n ID might be invalid, or the multipart upload might have been aborted or\n completed.\n
\n\n HTTP Status Code: 404 Not Found \n
\n\n SOAP Fault Code Prefix: Client\n
\n\n Related Resources\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part by copying data from an existing object as data source. You specify the\n data source by adding the request header x-amz-copy-source
in your request and\n a byte range by adding the request header x-amz-copy-source-range
in your\n request.
The minimum allowable part size for a multipart upload is 5 MB. For more information\n about multipart upload limits, go to Quick\n Facts in the Amazon S3 User Guide.
\nInstead of using an existing object as part data, you might use the UploadPart\n action and provide data in your request.
\nYou must initiate a multipart upload before you can upload any part. In response to your\n initiate request. Amazon S3 returns a unique identifier, the upload ID, that you must include in\n your upload part request.
\nFor more information about using the UploadPartCopy
operation, see the\n following:
For conceptual information about multipart uploads, see Uploading Objects Using Multipart\n Upload in the Amazon S3 User Guide.
\nFor information about permissions required to use the multipart upload API, see\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\nFor information about copying objects using a single atomic action vs. a multipart\n upload, see Operations on Objects in\n the Amazon S3 User Guide.
\nFor information about using server-side encryption with customer-provided\n encryption keys with the UploadPartCopy
operation, see CopyObject and UploadPart.
Note the following additional considerations about the request headers\n x-amz-copy-source-if-match
, x-amz-copy-source-if-none-match
,\n x-amz-copy-source-if-unmodified-since
, and\n x-amz-copy-source-if-modified-since
:
\n
\n Consideration 1 - If both of the\n x-amz-copy-source-if-match
and\n x-amz-copy-source-if-unmodified-since
headers are present in the\n request as follows:
\n x-amz-copy-source-if-match
condition evaluates to true
,\n and;
\n x-amz-copy-source-if-unmodified-since
condition evaluates to\n false
;
Amazon S3 returns 200 OK
and copies the data.\n
\n Consideration 2 - If both of the\n x-amz-copy-source-if-none-match
and\n x-amz-copy-source-if-modified-since
headers are present in the\n request as follows:
\n x-amz-copy-source-if-none-match
condition evaluates to\n false
, and;
\n x-amz-copy-source-if-modified-since
condition evaluates to\n true
;
Amazon S3 returns 412 Precondition Failed
response code.\n
\n Versioning\n
\nIf your bucket has versioning enabled, you could have multiple versions of the same\n object. By default, x-amz-copy-source
identifies the current version of the\n object to copy. If the current version is a delete marker and you don't specify a versionId\n in the x-amz-copy-source
, Amazon S3 returns a 404 error, because the object does\n not exist. If you specify versionId in the x-amz-copy-source
and the versionId\n is a delete marker, Amazon S3 returns an HTTP 400 error, because you are not allowed to specify\n a delete marker as a version for the x-amz-copy-source
.
You can optionally specify a specific version of the source object to copy by adding the\n versionId
subresource as shown in the following example:
\n x-amz-copy-source: /bucket/object?versionId=version id
\n
\n Special Errors\n
\n\n Code: NoSuchUpload\n
\n\n Cause: The specified multipart upload does not exist. The upload\n ID might be invalid, or the multipart upload might have been aborted or\n completed.\n
\n\n HTTP Status Code: 404 Not Found\n
\n\n Code: InvalidRequest\n
\n\n Cause: The specified copy source is not supported as a byte-range\n copy source.\n
\n\n HTTP Status Code: 400 Bad Request\n
\n\n Related Resources\n
\n\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part by copying data from an existing object as data source. You specify the\n data source by adding the request header x-amz-copy-source
in your request and\n a byte range by adding the request header x-amz-copy-source-range
in your\n request.
For information about maximum and minimum part sizes and other multipart upload specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nInstead of using an existing object as part data, you might use the UploadPart\n action and provide data in your request.
\nYou must initiate a multipart upload before you can upload any part. In response to your\n initiate request. Amazon S3 returns a unique identifier, the upload ID, that you must include in\n your upload part request.
\nFor more information about using the UploadPartCopy
operation, see the\n following:
For conceptual information about multipart uploads, see Uploading Objects Using Multipart\n Upload in the Amazon S3 User Guide.
\nFor information about permissions required to use the multipart upload API, see\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\nFor information about copying objects using a single atomic action vs. a multipart\n upload, see Operations on Objects in\n the Amazon S3 User Guide.
\nFor information about using server-side encryption with customer-provided\n encryption keys with the UploadPartCopy
operation, see CopyObject and UploadPart.
Note the following additional considerations about the request headers\n x-amz-copy-source-if-match
, x-amz-copy-source-if-none-match
,\n x-amz-copy-source-if-unmodified-since
, and\n x-amz-copy-source-if-modified-since
:
\n
\n Consideration 1 - If both of the\n x-amz-copy-source-if-match
and\n x-amz-copy-source-if-unmodified-since
headers are present in the\n request as follows:
\n x-amz-copy-source-if-match
condition evaluates to true
,\n and;
\n x-amz-copy-source-if-unmodified-since
condition evaluates to\n false
;
Amazon S3 returns 200 OK
and copies the data.\n
\n Consideration 2 - If both of the\n x-amz-copy-source-if-none-match
and\n x-amz-copy-source-if-modified-since
headers are present in the\n request as follows:
\n x-amz-copy-source-if-none-match
condition evaluates to\n false
, and;
\n x-amz-copy-source-if-modified-since
condition evaluates to\n true
;
Amazon S3 returns 412 Precondition Failed
response code.\n
\n Versioning\n
\nIf your bucket has versioning enabled, you could have multiple versions of the same\n object. By default, x-amz-copy-source
identifies the current version of the\n object to copy. If the current version is a delete marker and you don't specify a versionId\n in the x-amz-copy-source
, Amazon S3 returns a 404 error, because the object does\n not exist. If you specify versionId in the x-amz-copy-source
and the versionId\n is a delete marker, Amazon S3 returns an HTTP 400 error, because you are not allowed to specify\n a delete marker as a version for the x-amz-copy-source
.
You can optionally specify a specific version of the source object to copy by adding the\n versionId
subresource as shown in the following example:
\n x-amz-copy-source: /bucket/object?versionId=version id
\n
\n Special Errors\n
\n\n Code: NoSuchUpload\n
\n\n Cause: The specified multipart upload does not exist. The upload\n ID might be invalid, or the multipart upload might have been aborted or\n completed.\n
\n\n HTTP Status Code: 404 Not Found\n
\n\n Code: InvalidRequest\n
\n\n Cause: The specified copy source is not supported as a byte-range\n copy source.\n
\n\n HTTP Status Code: 400 Bad Request\n
\n\n Related Resources\n
\n\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\n