class Aws::S3::Types::CopyObjectRequest


@see docs.aws.amazon.com/goto/WebAPI/s3-2006-03-01/CopyObjectRequest AWS API Documentation
@return [String]
(access denied).
bucket, the request fails with the HTTP status code ‘403 Forbidden`
ID that you provide does not match the actual owner of the source
The account ID of the expected source bucket owner. If the account
@!attribute [rw] expected_source_bucket_owner
@return [String]
Forbidden` (access denied).
destination bucket, the request fails with the HTTP status code `403
account ID that you provide does not match the actual owner of the
The account ID of the expected destination bucket owner. If the
@!attribute [rw] expected_bucket_owner
@return [String]
</note>
<note markdown=“1”> This functionality is not supported for directory buckets.
Specifies whether you want to apply a legal hold to the object copy.
@!attribute [rw] object_lock_legal_hold_status
@return [Time]
</note>
<note markdown=“1”> This functionality is not supported for directory buckets.
to expire.
The date and time when you want the Object Lock of the object copy
@!attribute [rw] object_lock_retain_until_date
@return [String]
</note>
<note markdown=“1”> This functionality is not supported for directory buckets.
The Object Lock mode that you want to apply to the object copy.
@!attribute [rw] object_lock_mode
@return [String]
</note>
value of `x-amz-tagging` is the empty value.
directory bucket destination object. This is because the default
source object and don’t set the ‘x-amz-tagging` value of the
* When you attempt to `REPLACE` the tag-set of a directory bucket
to empty.
`x-amz-tagging` value of the directory bucket destination object
bucket source object that has non-empty tags and set the
* When you attempt to `REPLACE` the tag-set of a general purpose
bucket destination object to empty.
source object and set the `x-amz-tagging` value of the directory
* When you attempt to `REPLACE` the tag-set of a directory bucket
copies an empty tag-set to the destination object.
source object that has no tags to a general purpose bucket. It
* When you attempt to `COPY` the tag-set from a directory bucket
a `CopyObject` operation, the following situations are allowed:
Because only the empty tag-set is supported for directory buckets in
value of `x-amz-tagging-directive` is `COPY`.
source object has non-empty tags. This is because the default
* When you don’t set the ‘x-amz-tagging-directive` header and the
set a non-empty value to `x-amz-tagging`.
* When you attempt to `REPLACE` the tag-set of a source object and
that has non-empty tags.
* When you attempt to `COPY` the tag-set from an S3 source object
response in any of the following situations:
a directory bucket, you will receive a `501 Not Implemented`
a `501 Not Implemented` status code. When the destination bucket is
attempt to write non-empty tags into directory buckets will receive
operation, only the empty tag-set is supported. Any requests that
<note markdown=“1”> **Directory buckets** - For directory buckets in a `CopyObject`
The default value is the empty value.
Query parameters.
from the source object directly. The tag-set must be encoded as URL
set the `x-amz-tagging` header, because the tag-set will be copied
choose `COPY` for the `x-amz-tagging-directive`, you don’t need to
if you choose ‘REPLACE` for the `x-amz-tagging-directive`. If you
value must be used in conjunction with the `x-amz-tagging-directive`
The tag-set for the object copy in the destination bucket. This
@!attribute [rw] tagging
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/dev/ObjectsinRequesterPaysBuckets.html<br><br><br><br></note>
<note markdown=“1”> This functionality is not supported for directory buckets.
Requester Pays Buckets] in the *Amazon S3 User Guide*.
objects from Requester Pays buckets, see [Downloading Objects in
charges to copy the object. For information about downloading
Requester Pays enabled, the requester will pay for corresponding
requests. If either the source or destination S3 bucket has
request. Bucket owners need not specify this parameter in their
Confirms that the requester knows that they will be charged for the
@!attribute [rw] request_payer
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the source object is in a
request so that Amazon S3 can decrypt the object for copying.
SSE-C, you must provide the necessary encryption information in your
If the source object for the copy is stored in Amazon S3 using
to ensure that the encryption key was transmitted without error.
RFC 1321. Amazon S3 uses this header for a message integrity check
Specifies the 128-bit MD5 digest of the encryption key according to
@!attribute [rw] copy_source_sse_customer_key_md5
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the source object is in a
request so that Amazon S3 can decrypt the object for copying.
SSE-C, you must provide the necessary encryption information in your
If the source object for the copy is stored in Amazon S3 using
created.
header must be the same one that was used when the source object was
to decrypt the source object. The encryption key provided in this
Specifies the customer-provided encryption key for Amazon S3 to use
@!attribute [rw] copy_source_sse_customer_key
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the source object is in a
request so that Amazon S3 can decrypt the object for copying.
SSE-C, you must provide the necessary encryption information in your
If the source object for the copy is stored in Amazon S3 using
(for example, `AES256`).
Specifies the algorithm to use when decrypting the source object
@!attribute [rw] copy_source_sse_customer_algorithm
@return [Boolean]<br>: docs.aws.amazon.com/AmazonS3/latest/API/API_CopyObject.html<br>[1]: docs.aws.amazon.com/AmazonS3/latest/dev/bucket-key.html<br><br><br><br></note>
is made for a KMS-encrypted object.
this case, Amazon S3 makes a call to KMS every time a copy request
buckets, or between directory buckets, through [CopyObject]. In
directory buckets, from directory buckets to general purpose
copy SSE-KMS encrypted objects from general purpose buckets to
<note markdown=“1”> **Directory buckets** - S3 Bucket Keys aren’t supported, when you
S3 User Guide*.
For more information, see [Amazon S3 Bucket Keys] in the *Amazon
Key.
a COPY action doesn’t affect bucket-level settings for S3 Bucket
Key for object encryption with SSE-KMS. Specifying this header with
Setting this header to ‘true` causes Amazon S3 to use an S3 Bucket
enable an S3 Bucket Key for the object.
(KMS) keys (SSE-KMS). If a target object uses SSE-KMS, you can
encryption with server-side encryption using Key Management Service
Specifies whether Amazon S3 should use an S3 Bucket Key for object
@!attribute [rw] bucket_key_enabled
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/userguide/UsingKMSEncryption.html#encryption-context<br><br><br><br>additional encryption context value is not supported.
encryption context - the bucket Amazon Resource Name (ARN). An
encryption context value. The value must match the default
**Directory buckets** - You can optionally provide an explicit<br><br>context] in the *Amazon S3 User Guide*.
to the destination object. For more information, see [Encryption
additional encryption context of the source object won’t be copied
additional encryption context for your destination object. The
specify encryption context for ‘CopyObject` requests if you want an
**General purpose buckets** - This value must be explicitly added to
string holding JSON with the encryption context key-value pairs.
encryption. The value of this header is a base64-encoded UTF-8
additional encryption context to use for the destination object
Specifies the Amazon Web Services KMS Encryption Context as an
@!attribute [rw] ssekms_encryption_context
@return [String]<br>: docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk<br>[2]: docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk<br>[1]: docs.aws.amazon.com/AmazonS3/latest/dev/UsingAWSSDK.html#specify-signature-version<br><br><br><br>specification results in an HTTP `400 Bad Request` error.
Services managed key] (`aws/s3`) isn’t supported. Incorrect key
managed key] per directory bucket’s lifetime. The [Amazon Web
alias). Your SSE-KMS configuration can only support 1 [customer
the bucket’s default customer managed key (using key ID or ARN, not
x-amz-server-side-encryption-aws-kms-key-id` header, it must match
key ID. If you want to explicitly set the ‘
header implicitly uses the bucket’s default KMS customer managed
`aws:kms`. Then, the ‘x-amz-server-side-encryption-aws-kms-key-id`
recommended to specify the `x-amz-server-side-encryption` header to
**Directory buckets** - To encrypt data using SSE-KMS, it’s
User Guide*.
Signature Version in Request Authentication] in the *Amazon S3
Web Services SDKs and Amazon Web Services CLI, see [Specifying the
information about configuring any of the officially supported Amazon
by KMS will fail if they’re not made via SSL or using SigV4. For
object encryption. All GET and PUT requests for an object protected
Specifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for
@!attribute [rw] ssekms_key_id
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the destination bucket is a
to ensure that the encryption key was transmitted without error.
RFC 1321. Amazon S3 uses this header for a message integrity check
Specifies the 128-bit MD5 digest of the encryption key according to
@!attribute [rw] sse_customer_key_md5
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the destination bucket is a
`x-amz-server-side-encryption-customer-algorithm` header.
key must be appropriate for use with the algorithm specified in the
it is discarded. Amazon S3 does not store the encryption key. The
in encrypting data. This value is used to store the object and then
Specifies the customer-provided encryption key for Amazon S3 to use
@!attribute [rw] sse_customer_key
@return [String]
</note>
directory bucket.
<note markdown=“1”> This functionality is not supported when the destination bucket is a
precedence.
destination bucket, the encryption setting in your request takes
different from the default encryption configuration of the
customer-provided key. If the encryption setting in your request is
object with an Amazon S3 managed key, a KMS key, or a
specify appropriate encryption-related headers to encrypt the target
different type of encryption setting for the target object, you can
When you perform a ‘CopyObject` operation, if you want to use a
example, `AES256`).
Specifies the algorithm to use when encrypting the object (for
@!attribute [rw] sse_customer_algorithm
@return [String]
</note>
<note markdown=“1”> This functionality is not supported for directory buckets.
`x-amz-metadata-directive` header.
you may opt to provide this header in combination with the
copied when using the `x-amz-metadata-directive` header. Instead,
the object metadata. This value is unique to each object and is not
or to an external URL. Amazon S3 stores the value of this header in
requests for this object copy to another object in the same bucket
If the destination bucket is configured as a website, redirects
@!attribute [rw] website_redirect_location
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/dev/CopyingObjectsExamples.html<br>[3]: docs.aws.amazon.com/AmazonS3/latest/API/API_RestoreObject.html<br>[2]: docs.aws.amazon.com/AmazonS3/latest/userguide/intelligent-tiering-overview.html#intel-tiering-tier-definition<br>[1]: docs.aws.amazon.com/AmazonS3/latest/dev/storage-class-intro.html<br><br><br><br>Objects][4] in the *Amazon S3 User Guide*.
For more information, see [RestoreObject] and [Copying
Access` or `Deep Archive Access`.
and it’s [S3 Intelligent-Tiering access tier] is ‘Archive
* The storage class of the source object is `INTELLIGENT_TIERING`
`DEEP_ARCHIVE`.
* The storage class of the source object is `GLACIER` or
conditions:
you must restore a copy of it if it meets any of the following
Before using an object as a source object for the copy operation,<br><br>Classes] in the *Amazon S3 User Guide*.
`x-amz-storage-class` header. For more information, see [Storage
an object that is already stored in Amazon S3 by using the
You can use the `CopyObject` action to change the storage class of
</note>
`OUTPOSTS` Storage Class.
* Amazon S3 on Outposts - S3 on Outposts only uses the
respond with the HTTP status code `400 Bad Request`.
storage class values won’t write a destination object and will
Access storage class) in Dedicated Local Zones. Unsupported
Availability Zones and ‘ONEZONE_IA` (the S3 One Zone-Infrequent
`EXPRESS_ONEZONE` (the S3 Express One Zone storage class) in
<note markdown=“1”> * Directory buckets - Directory buckets only support
different Storage Class.
availability. Depending on performance needs, you can specify a
`STANDARD` storage class provides high durability and high
will be stored in the `STANDARD` Storage Class by default. The
If the `x-amz-storage-class` header is not used, the copied object
@!attribute [rw] storage_class
@return [String]<br>: docs.aws.amazon.com/kms/latest/developerguide/concepts.html#aws-managed-cmk<br>[4]: docs.aws.amazon.com/kms/latest/developerguide/concepts.html#customer-cmk<br>[3]: docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-specifying-kms-encryption.html<br>[2]: docs.aws.amazon.com/AmazonS3/latest/userguide/s3-express-serv-side-encryption.html<br>[1]: docs.aws.amazon.com/AmazonS3/latest/dev/serv-side-encryption.html<br><br><br><br>read. These processes are handled transparently by Amazon FSx.
written to the file system, and automatically decrypted as it is
encrypted at rest. Data is automatically encrypted before being
file systems have encryption configured by default and are
valid server side encryption option is `aws:fsx`. All Amazon FSx
stored in Amazon FSx file systems using S3 access points, the only
* S3 access points for Amazon FSx - When accessing data
directory bucket’s default encryption configuration.
key is the same customer managed key that you specified for the
encryption-related request headers, you must ensure the encryption
encryption settings for new object copies with SSE-KMS in the
perform a ‘CopyObject` operation and want to specify server-side
key for the bucket’s SSE-KMS configuration. Then, when you
managed key for SSE-KMS, you can’t override the customer managed
for the lifetime of the bucket. After you specify a customer
can only support 1 [customer managed key] per directory bucket<br>key] (‘aws/s3`) isn’t supported. Your SSE-KMS configuration
[customer managed key]). The [Amazon Web Services managed
default encryption configuration with a KMS key (specifically, a
we recommend you specify SSE-KMS as the directory bucket’s
* To encrypt new object copies to a directory bucket with SSE-KMS,
object uploads].
buckets, see [Specifying server-side encryption with KMS for new
information about the encryption overriding behaviors in directory
server-side encryption] in the *Amazon S3 User Guide*. For more
settings. For more information, see [Protecting data with
objects are automatically encrypted with the desired encryption
`CreateSession` requests or ‘PUT` object requests. Then, new
you don’t override the bucket default encryption in your
default encryption uses the desired encryption configuration and
KMS keys (SSE-KMS) (‘aws:kms`). We recommend that the bucket’s
managed keys (SSE-S3) (‘AES256`) and server-side encryption with
server-side encryption: server-side encryption with Amazon S3
* For directory buckets, there are only two supported options for
Directory buckets
precedence.
destination bucket, the encryption setting in your request takes
is different from the default encryption configuration of the
customer-provided key. If the encryption setting in your request
target object with an Amazon S3 managed key, a KMS key, or a
can specify appropriate encryption-related headers to encrypt the
different type of encryption setting for the target object, you
* When you perform a `CopyObject` operation, if you want to use a
or a customer-provided key to encrypt the target object copy.
encryption keys (SSE-C). Amazon S3 uses the corresponding KMS key,
(DSSE-KMS), and server-side encryption with customer-provided
server-side encryption with Amazon Web Services KMS keys
Key Management Service (KMS) keys (SSE-KMS), dual-layer
options for server-side encryption: server-side encryption with
* For general purpose buckets, there are the following supported
General purpose buckets
User Guide*.
encryption, see [Using Server-Side Encryption] in the *Amazon S3
when you access it. For more information about server-side
writes your data to disks in its data centers and decrypts the data
With server-side encryption, Amazon S3 encrypts your data as it
corresponding encryption key to encrypt the target object copy.
different default encryption configuration, Amazon S3 uses the
Amazon S3 managed keys (SSE-S3). If the destination bucket has a
of encryption configuration that uses server-side encryption with
of the destination bucket. By default, all buckets have a base level
of the target object is set to the default encryption configuration
encryption information in your copy request, the encryption setting
an S3 bucket. When copying an object, if you don’t specify
Amazon S3 automatically encrypts all new objects that are copied to
destination object and will receive a ‘400 Bad Request` response.
in Amazon S3. Unrecognized or unsupported values won’t write a
The server-side encryption algorithm used when storing this object
@!attribute [rw] server_side_encryption
@return [String]
</note>
value of `x-amz-tagging` is the empty value.
directory bucket destination object. This is because the default
source object and don’t set the ‘x-amz-tagging` value of the
* When you attempt to `REPLACE` the tag-set of a directory bucket
to empty.
`x-amz-tagging` value of the directory bucket destination object
bucket source object that has non-empty tags and set the
* When you attempt to `REPLACE` the tag-set of a general purpose
bucket destination object to empty.
source object and set the `x-amz-tagging` value of the directory
* When you attempt to `REPLACE` the tag-set of a directory bucket
copies an empty tag-set to the destination object.
source object that has no tags to a general purpose bucket. It
* When you attempt to `COPY` the tag-set from a directory bucket
a `CopyObject` operation, the following situations are allowed:
Because only the empty tag-set is supported for directory buckets in
value of `x-amz-tagging-directive` is `COPY`.
source object has non-empty tags. This is because the default
* When you don’t set the ‘x-amz-tagging-directive` header and the
set a non-empty value to `x-amz-tagging`.
* When you attempt to `REPLACE` the tag-set of a source object and
that has non-empty tags.
* When you attempt to `COPY` the tag-set from an S3 source object
response in any of the following situations:
a directory bucket, you will receive a `501 Not Implemented`
a `501 Not Implemented` status code. When the destination bucket is
attempt to write non-empty tags into directory buckets will receive
operation, only the empty tag-set is supported. Any requests that
<note markdown=“1”> **Directory buckets** - For directory buckets in a `CopyObject`
The default value is `COPY`.
object or replaced with the tag-set that’s provided in the request.
Specifies whether the object tag-set is copied from the source
@!attribute [rw] tagging_directive
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/dev/amazon-s3-policy-keys.html<br><br><br><br></note>
request header.
the value, you must specify ‘x-amz-website-redirect-location` in the
not copied when using the `x-amz-metadata-directive` header. To copy
<note markdown=“1”> `x-amz-website-redirect-location` is unique to each object and is<br><br>examples] in the *Amazon S3 User Guide*.
uploaded. For more information, see [Amazon S3 condition key
condition key to enforce certain metadata behavior when objects are
grant permissions, you can use the `s3:x-amz-metadata-directive`
**General purpose bucket** - For general purpose buckets, when you
behavior.
new metadata. If this header isn’t specified, `COPY` is the default
an object, you can preserve all metadata (the default) or specify
replaced with metadata that’s provided in the request. When copying
Specifies whether the metadata is copied from the source object or
@!attribute [rw] metadata_directive
@return [Hash<String,String>]
A map of metadata to store with the object in S3.
@!attribute [rw] metadata
@return [String]
The key of the destination object.
@!attribute [rw] key
@return [String]
</note>
* This functionality is not supported for Amazon S3 on Outposts.
<note markdown=“1”> * This functionality is not supported for directory buckets.
Allows grantee to write the ACL for the applicable object.
@!attribute [rw] grant_write_acp
@return [String]
</note>
* This functionality is not supported for Amazon S3 on Outposts.
<note markdown=“1”> * This functionality is not supported for directory buckets.
Allows grantee to read the object ACL.
@!attribute [rw] grant_read_acp
@return [String]
</note>
* This functionality is not supported for Amazon S3 on Outposts.
<note markdown=“1”> * This functionality is not supported for directory buckets.
Allows grantee to read the object data and its metadata.
@!attribute [rw] grant_read
@return [String]
</note>
* This functionality is not supported for Amazon S3 on Outposts.
<note markdown=“1”> * This functionality is not supported for directory buckets.
object.
Gives the grantee READ, READ_ACP, and WRITE_ACP permissions on the
@!attribute [rw] grant_full_control
@return [Time]
The date and time at which the object is no longer cacheable.
@!attribute [rw] expires
@return [Time]
false
* ‘x-amz-copy-source-if-unmodified-since` condition evaluates to
* `x-amz-copy-source-if-match` condition evaluates to true
copies the data:
request and evaluate as follows, Amazon S3 returns `200 OK` and
`x-amz-copy-source-if-unmodified-since` headers are present in the
If both the `x-amz-copy-source-if-match` and
time.
Copies the object if it hasn’t been modified since the specified
@!attribute [rw] copy_source_if_unmodified_since
@return [String]
* ‘x-amz-copy-source-if-modified-since` condition evaluates to true
* `x-amz-copy-source-if-none-match` condition evaluates to false
Precondition Failed` response code:
request and evaluate as follows, Amazon S3 returns the `412
`x-amz-copy-source-if-modified-since` headers are present in the
If both the `x-amz-copy-source-if-none-match` and
specified ETag.
Copies the object if its entity tag (ETag) is different than the
@!attribute [rw] copy_source_if_none_match
@return [Time]
* `x-amz-copy-source-if-modified-since` condition evaluates to true
* `x-amz-copy-source-if-none-match` condition evaluates to false
Precondition Failed` response code:
request and evaluate as follows, Amazon S3 returns the `412
`x-amz-copy-source-if-modified-since` headers are present in the
If both the `x-amz-copy-source-if-none-match` and
Copies the object if it has been modified since the specified time.
@!attribute [rw] copy_source_if_modified_since
@return [String]
false
* `x-amz-copy-source-if-unmodified-since` condition evaluates to
* `x-amz-copy-source-if-match` condition evaluates to true
copies the data:
request and evaluate as follows, Amazon S3 returns `200 OK` and
`x-amz-copy-source-if-unmodified-since` headers are present in the
If both the `x-amz-copy-source-if-match` and
tag.
Copies the object if its entity tag (ETag) matches the specified
@!attribute [rw] copy_source_if_match
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/userguide/access-points.html<br><br><br><br></note>
for directory buckets.
<note markdown=“1”> **Directory buckets** - S3 Versioning isn’t enabled and supported
`x-amz-version-id` response header is always null.
bucket, the version ID that Amazon S3 generates in the
If you do not enable versioning or suspend it on the destination
`x-amz-version-id` response header in the response.
returns the version ID of the copied object in the
is different from the version ID of the source object. Amazon S3
generates a unique version ID for the copied object. This version ID
If you enable versioning on the destination bucket, Amazon S3
version of the source object.
If you don’t specify a version ID, Amazon S3 copies the latest
`awsexamplebucket/reports/january.pdf?versionId=QUpfdndhfd8438MNFDN93jdnJFkdmqnh893`).
‘?versionId=<version-id>` to the value (for example,
`versionId` query parameter. Specifically, append
as if the object was deleted. To copy a different version, use the
copy. If the current version is a delete marker, Amazon S3 behaves
header by default identifies the current version of an object to
If your source bucket versioning is enabled, the `x-amz-copy-source`
The value must be URL-encoded.
`arn:aws:s3-outposts:us-west-2:123456789012:outpost/my-outpost/object/reports/january.pdf`.
`us-west-2`, use the URL encoding of
outpost `my-outpost` owned by account `123456789012` in Region
For example, to copy the object `reports/january.pdf` through
`arn:aws:s3-outposts:<Region>:<account-id>:outpost/<outpost-id>/object/<key>`.
specify the ARN of the object as accessed in the format
Alternatively, for objects accessed through Amazon S3 on Outposts,
</note>
* Access points are not supported by directory buckets.
Services Region.
the source and destination buckets are in the same Amazon Web
<note markdown=“1”> * Amazon S3 supports copy operations using Access points only when
The value must be URL encoded.
`arn:aws:s3:us-west-2:123456789012:accesspoint/my-access-point/object/reports/january.pdf`.
Region `us-west-2`, use the URL encoding of
access point `my-access-point` owned by account `123456789012` in
For example, to copy the object `reports/january.pdf` through
`arn:aws:s3:<Region>:<account-id>:accesspoint/<access-point-name>/object/<key>`.
point, in the format
Resource Name (ARN) of the object as accessed through the access
* For objects accessed through access points, specify the Amazon
must be URL-encoded.
`awsexamplebucket–use1-az5–x-s3/reports/january.pdf`. The value
`awsexamplebucket–use1-az5–x-s3`, use
`reports/january.pdf` from the directory bucket
The value must be URL-encoded. To copy the object
`awsexamplebucket`, use `awsexamplebucket/reports/january.pdf`.
`reports/january.pdf` from the general purpose bucket
by a slash (/). For example, to copy the object
of the source bucket and the key of the source object, separated
* For objects not accessed through an access point, specify the name
[access point]:
depending on whether you want to access the source object through an
You specify the value of the copy source in one of two formats,
destination bucket.
single part object after the source object is copied to the
uploaded by using a multipart upload, the object copy will be a
object can be up to 5 GB. If the source object is an object that was
Specifies the source object for the copy operation. The source
@!attribute [rw] copy_source
@return [String]
A standard MIME type that describes the format of the object data.
@!attribute [rw] content_type
@return [String]
The language the content is in.
@!attribute [rw] content_language
@return [String]
</note>
this header field.
<note markdown=“1”> For directory buckets, only the `aws-chunked` value is supported in
media-type referenced by the Content-Type header field.
thus what decoding mechanisms must be applied to obtain the
Specifies what content encodings have been applied to the object and
@!attribute [rw] content_encoding
@return [String]
downloaded file.
as a file. It allows specifying the desired filename for the
whether an object should be displayed in a web browser or downloaded
Specifies presentational information for the object. Indicates
@!attribute [rw] content_disposition
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/userguide/checking-object-integrity.html<br><br><br><br></note>
performance.
`CRC32` is the default checksum algorithm that’s used for
<note markdown=“1”> For directory buckets, when you use Amazon Web Services SDKs,
values will respond with the HTTP status code ‘400 Bad Request`.
`x-amz-checksum-algorithm` header. Unrecognized or unsupported
different checksum algorithm to use with the
present on the source object). You can optionally specify a
copied from the source object to the destination object (if it’s
`x-amz-checksum-algorithm` header, the checksum algorithm will be
`CopyObject` request does not include this
checksum value will be copied to the new object by default. If the
When you copy an object, if the source object has a checksum, that<br><br>integrity] in the *Amazon S3 User Guide*.
checksum for the object. For more information, see [Checking object
Indicates the algorithm that you want Amazon S3 to use to create the
@!attribute [rw] checksum_algorithm
@return [String]
Specifies the caching behavior along the request/reply chain.
@!attribute [rw] cache_control
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/userguide/S3onOutposts.html<br>[2]: docs.aws.amazon.com/AmazonS3/latest/userguide/using-access-points.html<br>[1]: docs.aws.amazon.com/AmazonS3/latest/userguide/directory-bucket-naming-rules.html<br><br><br><br>CLI or SDKs.
The hostname isn’t required when you use the Amazon Web Services
AccessPointName-AccountId.outpostID.s3-outposts.Region.amazonaws.com`.
requests to the S3 on Outposts hostname, in the format ‘
action with S3 on Outposts through the REST API, you must direct<br>Outposts?] in the *S3 on Outposts guide*. When you use this
For more information about S3 on Outposts, see [What is S3 on
Outposts, or between Outposts buckets and any other bucket types.
different Amazon Web Services Outposts, between buckets on the same
the same Outpost bucket. It’s not supported to copy objects across
alias for the destination bucket. You can only copy objects within
you must use the Outpost bucket access point ARN or the access point
**S3 on Outposts** - When you use this action with S3 on Outposts,
</note>
<note markdown=“1”> Object Lambda access points are not supported by directory buckets.
[Using access points] in the *Amazon S3 User Guide*.
bucket name. For more information about access point ARNs, see
Services SDKs, you provide the access point ARN in place of the
When using this action with an access point through the Amazon Web
AccessPointName-AccountId.s3-accesspoint.Region.amazonaws.com.
takes the form
requests to the access point hostname. The access point hostname
bucket name. When using the access point ARN, you must direct
buckets, you must provide the access point name in place of the
ARN. When you use this action with an access point for directory
access point in place of the bucket name or specify the access point
for general purpose buckets, you must provide the alias of the
**Access points** - When you use this action with an access point
</note>
HTTP ‘400 Bad Request` error with the error code `InvalidRequest`.
the same parent Amazon Web Services Region. Otherwise, you get an
Services Local Zones. The source and destination buckets must have
supported when the source or destination bucket is in Amazon Web
<note markdown=“1”> Copying objects across different Amazon Web Services Regions isn’t
*Amazon S3 User Guide*.
naming restrictions, see [Directory bucket naming rules] in the
amzn-s3-demo-bucket–usw2-az1–x-s3`). For information about bucket
follow the format ‘ bucket-base-name–zone-id–x-s3` (for example, `
the chosen Zone (Availability Zone or Local Zone). Bucket names must
requests are not supported. Directory bucket names must be unique in
Bucket-name.s3express-zone-id.region-code.amazonaws.com`. Path-style
bucket, you must use virtual-hosted-style requests in the format `
**Directory buckets** - When you use this operation with a directory
The name of the destination bucket.
@!attribute [rw] bucket
@return [String]<br>: docs.aws.amazon.com/AmazonS3/latest/userguide/about-object-ownership.html<br>[1]: docs.aws.amazon.com/AmazonS3/latest/dev/S3_ACLs_UsingACLs.html<br><br><br><br></note>
* This functionality is not supported for Amazon S3 on Outposts.
* This functionality is not supported for directory buckets.
account will be owned by the bucket owner.
for Object Ownership, all objects written to the bucket by any
<note markdown=“1”> * If your destination bucket uses the bucket owner enforced setting
*Amazon S3 User Guide*.
[Controlling ownership of objects and disabling ACLs] in the
this ACL expressed in the XML format. For more information, see
the `bucket-owner-full-control` canned ACL or an equivalent form of
`PUT` requests that specify bucket owner full control ACLs, such as
setting only accept `PUT` requests that don’t specify an ACL or
disabled and no longer affect permissions. Buckets that use this
bucket owner enforced setting for S3 Object Ownership, ACLs are
If the destination bucket that you’re copying objects to uses the
generate a copy request. For more information, see [Using ACLs].
To override the default ACL setting, specify a new ACL when you
set to ‘private` by default. Only the owner has full access control.
When you copy an object, the ACL metadata is not preserved and is
The canned access control list (ACL) to apply to the object.
@!attribute [rw] acl