Struct aws_sdk_dynamodb::client::fluent_builders::BatchWriteItem [−][src]
pub struct BatchWriteItem<C = DynConnector, M = AwsMiddleware, R = Standard> { /* fields omitted */ }
Expand description
Fluent builder constructing a request to BatchWriteItem
.
The BatchWriteItem
operation puts or deletes multiple items in one or more
tables. A single call to BatchWriteItem
can write up to 16 MB of data,
which can comprise as many as 25 put or delete requests. Individual items to be written
can be as large as 400 KB.
BatchWriteItem
cannot update items. To update items, use the UpdateItem
action.
The individual PutItem
and DeleteItem
operations specified in
BatchWriteItem
are atomic; however BatchWriteItem
as a whole is not. If any
requested operations fail because the table's provisioned throughput is exceeded or an
internal processing failure occurs, the failed operations are returned in the
UnprocessedItems
response parameter. You can investigate and optionally resend the
requests. Typically, you would call BatchWriteItem
in a loop. Each iteration would
check for unprocessed items and submit a new BatchWriteItem
request with those
unprocessed items until all items have been processed.
If none of the items can be processed due to insufficient
provisioned throughput on all of the tables in the request, then
BatchWriteItem
returns a
ProvisionedThroughputExceededException
.
If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.
For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide.
With BatchWriteItem
, you can efficiently write or delete large amounts of
data, such as from Amazon EMR, or copy data from another database into DynamoDB. In
order to improve performance with these large-scale operations,
BatchWriteItem
does not behave in the same way as individual
PutItem
and DeleteItem
calls would. For example, you
cannot specify conditions on individual put and delete requests, and
BatchWriteItem
does not return deleted items in the response.
If you use a programming language that supports concurrency, you can use
threads to write items in parallel. Your application must include the necessary logic to
manage the threads. With languages that don't support threading, you must update
or delete the specified items one at a time. In both situations, BatchWriteItem
performs the specified put and delete operations in
parallel, giving you the power of the thread pool approach without having to introduce
complexity into your application.
Parallel processing reduces latency, but each specified put and delete request consumes the same number of write capacity units whether it is processed in parallel or not. Delete operations on nonexistent items consume one write capacity unit.
If one or more of the following is true, DynamoDB rejects the entire batch write operation:
-
One or more tables specified in the
BatchWriteItem
request does not exist. -
Primary key attributes specified on an item in the request do not match those in the corresponding table's primary key schema.
-
You try to perform multiple operations on the same item in the same
BatchWriteItem
request. For example, you cannot put and delete the same item in the sameBatchWriteItem
request. -
Your request contains at least two items with identical hash and range keys (which essentially is two put operations).
-
There are more than 25 requests in the batch.
-
Any individual item in a batch exceeds 400 KB.
-
The total request size exceeds 16 MB.
Implementations
impl<C, M, R> BatchWriteItem<C, M, R> where
C: SmithyConnector,
M: SmithyMiddleware<C>,
R: NewRequestPolicy,
impl<C, M, R> BatchWriteItem<C, M, R> where
C: SmithyConnector,
M: SmithyMiddleware<C>,
R: NewRequestPolicy,
pub async fn send(
self
) -> Result<BatchWriteItemOutput, SdkError<BatchWriteItemError>> where
R::Policy: SmithyRetryPolicy<BatchWriteItemInputOperationOutputAlias, BatchWriteItemOutput, BatchWriteItemError, BatchWriteItemInputOperationRetryAlias>,
pub async fn send(
self
) -> Result<BatchWriteItemOutput, SdkError<BatchWriteItemError>> where
R::Policy: SmithyRetryPolicy<BatchWriteItemInputOperationOutputAlias, BatchWriteItemOutput, BatchWriteItemError, BatchWriteItemInputOperationRetryAlias>,
Sends the request and returns the response.
If an error occurs, an SdkError
will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
Adds a key-value pair to RequestItems
.
To override the contents of this collection use set_request_items
.
A map of one or more table names and, for each table, a list of operations to be performed
(DeleteRequest
or PutRequest
). Each element in the map consists of the
following:
-
DeleteRequest
- Perform aDeleteItem
operation on the specified item. The item to be deleted is identified by aKey
subelement:-
Key
- A map of primary key attribute values that uniquely identify the item. Each entry in this map consists of an attribute name and an attribute value. For each primary key, you must provide all of the key attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.
-
-
PutRequest
- Perform aPutItem
operation on the specified item. The item to be put is identified by anItem
subelement:-
Item
- A map of attributes and their values. Each entry in this map consists of an attribute name and an attribute value. Attribute values must not be null; string and binary type attributes must have lengths greater than zero; and set type attributes must not be empty. Requests that contain empty values are rejected with aValidationException
exception.If you specify any attributes that are part of an index key, then the data types for those attributes must match those of the schema in the table's attribute definition.
-
A map of one or more table names and, for each table, a list of operations to be performed
(DeleteRequest
or PutRequest
). Each element in the map consists of the
following:
-
DeleteRequest
- Perform aDeleteItem
operation on the specified item. The item to be deleted is identified by aKey
subelement:-
Key
- A map of primary key attribute values that uniquely identify the item. Each entry in this map consists of an attribute name and an attribute value. For each primary key, you must provide all of the key attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.
-
-
PutRequest
- Perform aPutItem
operation on the specified item. The item to be put is identified by anItem
subelement:-
Item
- A map of attributes and their values. Each entry in this map consists of an attribute name and an attribute value. Attribute values must not be null; string and binary type attributes must have lengths greater than zero; and set type attributes must not be empty. Requests that contain empty values are rejected with aValidationException
exception.If you specify any attributes that are part of an index key, then the data types for those attributes must match those of the schema in the table's attribute definition.
-
Determines the level of detail about provisioned throughput consumption that is returned in the response:
-
INDEXES
- The response includes the aggregateConsumedCapacity
for the operation, together withConsumedCapacity
for each table and secondary index that was accessed.Note that some operations, such as
GetItem
andBatchGetItem
, do not access any indexes at all. In these cases, specifyingINDEXES
will only returnConsumedCapacity
information for table(s). -
TOTAL
- The response includes only the aggregateConsumedCapacity
for the operation. -
NONE
- NoConsumedCapacity
details are included in the response.
Determines the level of detail about provisioned throughput consumption that is returned in the response:
-
INDEXES
- The response includes the aggregateConsumedCapacity
for the operation, together withConsumedCapacity
for each table and secondary index that was accessed.Note that some operations, such as
GetItem
andBatchGetItem
, do not access any indexes at all. In these cases, specifyingINDEXES
will only returnConsumedCapacity
information for table(s). -
TOTAL
- The response includes only the aggregateConsumedCapacity
for the operation. -
NONE
- NoConsumedCapacity
details are included in the response.
Determines whether item collection metrics are returned. If set to SIZE
, the response includes statistics about item collections, if any, that were modified during
the operation are returned in the response. If set to NONE
(the default), no statistics are returned.
pub fn set_return_item_collection_metrics(
self,
input: Option<ReturnItemCollectionMetrics>
) -> Self
pub fn set_return_item_collection_metrics(
self,
input: Option<ReturnItemCollectionMetrics>
) -> Self
Determines whether item collection metrics are returned. If set to SIZE
, the response includes statistics about item collections, if any, that were modified during
the operation are returned in the response. If set to NONE
(the default), no statistics are returned.
Trait Implementations
Auto Trait Implementations
impl<C = DynConnector, M = AwsMiddleware, R = Standard> !RefUnwindSafe for BatchWriteItem<C, M, R>
impl<C, M, R> Send for BatchWriteItem<C, M, R> where
C: Send + Sync,
M: Send + Sync,
R: Send + Sync,
impl<C, M, R> Sync for BatchWriteItem<C, M, R> where
C: Send + Sync,
M: Send + Sync,
R: Send + Sync,
impl<C, M, R> Unpin for BatchWriteItem<C, M, R>
impl<C = DynConnector, M = AwsMiddleware, R = Standard> !UnwindSafe for BatchWriteItem<C, M, R>
Blanket Implementations
Mutably borrows from an owned value. Read more
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more