Struct aws_sdk_kinesis::client::fluent_builders::PutRecord
source · [−]pub struct PutRecord { /* private fields */ }
Expand description
Fluent builder constructing a request to PutRecord
.
Writes a single data record into an Amazon Kinesis data stream. Call PutRecord
to send data into the stream for real-time ingestion and subsequent processing, one record at a time. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second.
You must specify the name of the stream that captures, stores, and transports the data; a partition key; and the data blob itself.
The data blob can be any type of data; for example, a segment from a log file, geographic/location data, website clickstream data, and so on.
The partition key is used by Kinesis Data Streams to distribute data across shards. Kinesis Data Streams segregates the data records that belong to a stream into multiple shards, using the partition key associated with each data record to determine the shard to which a given data record belongs.
Partition keys are Unicode strings, with a maximum length limit of 256 characters for each key. An MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards using the hash key ranges of the shards. You can override hashing the partition key to determine the shard by explicitly specifying a hash value using the ExplicitHashKey
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
PutRecord
returns the shard ID of where the data record was placed and the sequence number that was assigned to the data record.
Sequence numbers increase over time and are specific to a shard within a stream, not across all shards within a stream. To guarantee strictly increasing ordering, write serially to a shard and use the SequenceNumberForOrdering
parameter. For more information, see Adding Data to a Stream in the Amazon Kinesis Data Streams Developer Guide.
After you write a record to a stream, you cannot modify that record or its order within the stream.
If a PutRecord
request cannot be processed because of insufficient provisioned throughput on the shard involved in the request, PutRecord
throws ProvisionedThroughputExceededException
.
By default, data records are accessible for 24 hours from the time that they are added to a stream. You can use IncreaseStreamRetentionPeriod
or DecreaseStreamRetentionPeriod
to modify this retention period.
Implementations
sourceimpl PutRecord
impl PutRecord
sourcepub async fn send(self) -> Result<PutRecordOutput, SdkError<PutRecordError>>
pub async fn send(self) -> Result<PutRecordOutput, SdkError<PutRecordError>>
Sends the request and returns the response.
If an error occurs, an SdkError
will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
sourcepub fn stream_name(self, input: impl Into<String>) -> Self
pub fn stream_name(self, input: impl Into<String>) -> Self
The name of the stream to put the data record into.
sourcepub fn set_stream_name(self, input: Option<String>) -> Self
pub fn set_stream_name(self, input: Option<String>) -> Self
The name of the stream to put the data record into.
sourcepub fn data(self, input: Blob) -> Self
pub fn data(self, input: Blob) -> Self
The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).
sourcepub fn set_data(self, input: Option<Blob>) -> Self
pub fn set_data(self, input: Option<Blob>) -> Self
The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).
sourcepub fn partition_key(self, input: impl Into<String>) -> Self
pub fn partition_key(self, input: impl Into<String>) -> Self
Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.
sourcepub fn set_partition_key(self, input: Option<String>) -> Self
pub fn set_partition_key(self, input: Option<String>) -> Self
Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.
sourcepub fn explicit_hash_key(self, input: impl Into<String>) -> Self
pub fn explicit_hash_key(self, input: impl Into<String>) -> Self
The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.
sourcepub fn set_explicit_hash_key(self, input: Option<String>) -> Self
pub fn set_explicit_hash_key(self, input: Option<String>) -> Self
The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.
sourcepub fn sequence_number_for_ordering(self, input: impl Into<String>) -> Self
pub fn sequence_number_for_ordering(self, input: impl Into<String>) -> Self
Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering
of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.
sourcepub fn set_sequence_number_for_ordering(self, input: Option<String>) -> Self
pub fn set_sequence_number_for_ordering(self, input: Option<String>) -> Self
Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering
of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.
Trait Implementations
Auto Trait Implementations
impl !RefUnwindSafe for PutRecord
impl Send for PutRecord
impl Sync for PutRecord
impl Unpin for PutRecord
impl !UnwindSafe for PutRecord
Blanket Implementations
sourceimpl<T> BorrowMut<T> for T where
T: ?Sized,
impl<T> BorrowMut<T> for T where
T: ?Sized,
const: unstable · sourcefn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Mutably borrows from an owned value. Read more
sourceimpl<T> Instrument for T
impl<T> Instrument for T
sourcefn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
sourcefn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
sourceimpl<T> ToOwned for T where
T: Clone,
impl<T> ToOwned for T where
T: Clone,
type Owned = T
type Owned = T
The resulting type after obtaining ownership.
sourcefn clone_into(&self, target: &mut T)
fn clone_into(&self, target: &mut T)
toowned_clone_into
)Uses borrowed data to replace owned data, usually by cloning. Read more
sourceimpl<T> WithSubscriber for T
impl<T> WithSubscriber for T
sourcefn with_subscriber<S>(self, subscriber: S) -> WithDispatch<Self> where
S: Into<Dispatch>,
fn with_subscriber<S>(self, subscriber: S) -> WithDispatch<Self> where
S: Into<Dispatch>,
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
sourcefn with_current_subscriber(self) -> WithDispatch<Self>
fn with_current_subscriber(self) -> WithDispatch<Self>
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more