pub struct CreateStreamProcessorFluentBuilder { /* private fields */ }
Expand description
Fluent builder constructing a request to CreateStreamProcessor
.
Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces or to detect labels in a streaming video.
Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. There are two different settings for stream processors in Amazon Rekognition: detecting faces and detecting labels.
-
If you are creating a stream processor for detecting faces, you provide as input a Kinesis video stream (
Input
) and a Kinesis data stream (Output
) stream for receiving the output. You must use theFaceSearch
option inSettings
, specifying the collection that contains the faces you want to recognize. After you have finished analyzing a streaming video, useStopStreamProcessor
to stop processing. -
If you are creating a stream processor to detect labels, you provide as input a Kinesis video stream (
Input
), Amazon S3 bucket information (Output
), and an Amazon SNS topic ARN (NotificationChannel
). You can also provide a KMS key ID to encrypt the data sent to your Amazon S3 bucket. You specify what you want to detect by using theConnectedHome
option in settings, and selecting one of the following:PERSON
,PET
,PACKAGE
,ALL
You can also specify where in the frame you want Amazon Rekognition to monitor withRegionsOfInterest
. When you run theStartStreamProcessor
operation on a label detection stream processor, you input start and stop information to determine the length of the processing time.
Use Name
to assign an identifier for the stream processor. You use Name
to manage the stream processor. For example, you can start processing the source video by calling StartStreamProcessor
with the Name
field.
This operation requires permissions to perform the rekognition:CreateStreamProcessor
action. If you want to tag your stream processor, you also require permission to perform the rekognition:TagResource
operation.
Implementations§
Source§impl CreateStreamProcessorFluentBuilder
impl CreateStreamProcessorFluentBuilder
Sourcepub fn as_input(&self) -> &CreateStreamProcessorInputBuilder
pub fn as_input(&self) -> &CreateStreamProcessorInputBuilder
Access the CreateStreamProcessor as a reference.
Sourcepub async fn send(
self,
) -> Result<CreateStreamProcessorOutput, SdkError<CreateStreamProcessorError, HttpResponse>>
pub async fn send( self, ) -> Result<CreateStreamProcessorOutput, SdkError<CreateStreamProcessorError, HttpResponse>>
Sends the request and returns the response.
If an error occurs, an SdkError
will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
Sourcepub fn customize(
self,
) -> CustomizableOperation<CreateStreamProcessorOutput, CreateStreamProcessorError, Self>
pub fn customize( self, ) -> CustomizableOperation<CreateStreamProcessorOutput, CreateStreamProcessorError, Self>
Consumes this builder, creating a customizable operation that can be modified before being sent.
Sourcepub fn input(self, input: StreamProcessorInput) -> Self
pub fn input(self, input: StreamProcessorInput) -> Self
Kinesis video stream stream that provides the source streaming video. If you are using the AWS CLI, the parameter name is StreamProcessorInput
. This is required for both face search and label detection stream processors.
Sourcepub fn set_input(self, input: Option<StreamProcessorInput>) -> Self
pub fn set_input(self, input: Option<StreamProcessorInput>) -> Self
Kinesis video stream stream that provides the source streaming video. If you are using the AWS CLI, the parameter name is StreamProcessorInput
. This is required for both face search and label detection stream processors.
Sourcepub fn get_input(&self) -> &Option<StreamProcessorInput>
pub fn get_input(&self) -> &Option<StreamProcessorInput>
Kinesis video stream stream that provides the source streaming video. If you are using the AWS CLI, the parameter name is StreamProcessorInput
. This is required for both face search and label detection stream processors.
Sourcepub fn output(self, input: StreamProcessorOutput) -> Self
pub fn output(self, input: StreamProcessorOutput) -> Self
Kinesis data stream stream or Amazon S3 bucket location to which Amazon Rekognition Video puts the analysis results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput
. This must be a S3Destination
of an Amazon S3 bucket that you own for a label detection stream processor or a Kinesis data stream ARN for a face search stream processor.
Sourcepub fn set_output(self, input: Option<StreamProcessorOutput>) -> Self
pub fn set_output(self, input: Option<StreamProcessorOutput>) -> Self
Kinesis data stream stream or Amazon S3 bucket location to which Amazon Rekognition Video puts the analysis results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput
. This must be a S3Destination
of an Amazon S3 bucket that you own for a label detection stream processor or a Kinesis data stream ARN for a face search stream processor.
Sourcepub fn get_output(&self) -> &Option<StreamProcessorOutput>
pub fn get_output(&self) -> &Option<StreamProcessorOutput>
Kinesis data stream stream or Amazon S3 bucket location to which Amazon Rekognition Video puts the analysis results. If you are using the AWS CLI, the parameter name is StreamProcessorOutput
. This must be a S3Destination
of an Amazon S3 bucket that you own for a label detection stream processor or a Kinesis data stream ARN for a face search stream processor.
Sourcepub fn name(self, input: impl Into<String>) -> Self
pub fn name(self, input: impl Into<String>) -> Self
An identifier you assign to the stream processor. You can use Name
to manage the stream processor. For example, you can get the current status of the stream processor by calling DescribeStreamProcessor
. Name
is idempotent. This is required for both face search and label detection stream processors.
Sourcepub fn set_name(self, input: Option<String>) -> Self
pub fn set_name(self, input: Option<String>) -> Self
An identifier you assign to the stream processor. You can use Name
to manage the stream processor. For example, you can get the current status of the stream processor by calling DescribeStreamProcessor
. Name
is idempotent. This is required for both face search and label detection stream processors.
Sourcepub fn get_name(&self) -> &Option<String>
pub fn get_name(&self) -> &Option<String>
An identifier you assign to the stream processor. You can use Name
to manage the stream processor. For example, you can get the current status of the stream processor by calling DescribeStreamProcessor
. Name
is idempotent. This is required for both face search and label detection stream processors.
Sourcepub fn settings(self, input: StreamProcessorSettings) -> Self
pub fn settings(self, input: StreamProcessorSettings) -> Self
Input parameters used in a streaming video analyzed by a stream processor. You can use FaceSearch
to recognize faces in a streaming video, or you can use ConnectedHome
to detect labels.
Sourcepub fn set_settings(self, input: Option<StreamProcessorSettings>) -> Self
pub fn set_settings(self, input: Option<StreamProcessorSettings>) -> Self
Input parameters used in a streaming video analyzed by a stream processor. You can use FaceSearch
to recognize faces in a streaming video, or you can use ConnectedHome
to detect labels.
Sourcepub fn get_settings(&self) -> &Option<StreamProcessorSettings>
pub fn get_settings(&self) -> &Option<StreamProcessorSettings>
Input parameters used in a streaming video analyzed by a stream processor. You can use FaceSearch
to recognize faces in a streaming video, or you can use ConnectedHome
to detect labels.
Sourcepub fn role_arn(self, input: impl Into<String>) -> Self
pub fn role_arn(self, input: impl Into<String>) -> Self
The Amazon Resource Number (ARN) of the IAM role that allows access to the stream processor. The IAM role provides Rekognition read permissions for a Kinesis stream. It also provides write permissions to an Amazon S3 bucket and Amazon Simple Notification Service topic for a label detection stream processor. This is required for both face search and label detection stream processors.
Sourcepub fn set_role_arn(self, input: Option<String>) -> Self
pub fn set_role_arn(self, input: Option<String>) -> Self
The Amazon Resource Number (ARN) of the IAM role that allows access to the stream processor. The IAM role provides Rekognition read permissions for a Kinesis stream. It also provides write permissions to an Amazon S3 bucket and Amazon Simple Notification Service topic for a label detection stream processor. This is required for both face search and label detection stream processors.
Sourcepub fn get_role_arn(&self) -> &Option<String>
pub fn get_role_arn(&self) -> &Option<String>
The Amazon Resource Number (ARN) of the IAM role that allows access to the stream processor. The IAM role provides Rekognition read permissions for a Kinesis stream. It also provides write permissions to an Amazon S3 bucket and Amazon Simple Notification Service topic for a label detection stream processor. This is required for both face search and label detection stream processors.
Adds a key-value pair to Tags
.
To override the contents of this collection use set_tags
.
A set of tags (key-value pairs) that you want to attach to the stream processor.
A set of tags (key-value pairs) that you want to attach to the stream processor.
A set of tags (key-value pairs) that you want to attach to the stream processor.
Sourcepub fn notification_channel(
self,
input: StreamProcessorNotificationChannel,
) -> Self
pub fn notification_channel( self, input: StreamProcessorNotificationChannel, ) -> Self
The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the object detection results and completion status of a video analysis operation.
Amazon Rekognition publishes a notification the first time an object of interest or a person is detected in the video stream. For example, if Amazon Rekognition detects a person at second 2, a pet at second 4, and a person again at second 5, Amazon Rekognition sends 2 object class detected notifications, one for a person at second 2 and one for a pet at second 4.
Amazon Rekognition also publishes an an end-of-session notification with a summary when the stream processing session is complete.
Sourcepub fn set_notification_channel(
self,
input: Option<StreamProcessorNotificationChannel>,
) -> Self
pub fn set_notification_channel( self, input: Option<StreamProcessorNotificationChannel>, ) -> Self
The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the object detection results and completion status of a video analysis operation.
Amazon Rekognition publishes a notification the first time an object of interest or a person is detected in the video stream. For example, if Amazon Rekognition detects a person at second 2, a pet at second 4, and a person again at second 5, Amazon Rekognition sends 2 object class detected notifications, one for a person at second 2 and one for a pet at second 4.
Amazon Rekognition also publishes an an end-of-session notification with a summary when the stream processing session is complete.
Sourcepub fn get_notification_channel(
&self,
) -> &Option<StreamProcessorNotificationChannel>
pub fn get_notification_channel( &self, ) -> &Option<StreamProcessorNotificationChannel>
The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the object detection results and completion status of a video analysis operation.
Amazon Rekognition publishes a notification the first time an object of interest or a person is detected in the video stream. For example, if Amazon Rekognition detects a person at second 2, a pet at second 4, and a person again at second 5, Amazon Rekognition sends 2 object class detected notifications, one for a person at second 2 and one for a pet at second 4.
Amazon Rekognition also publishes an an end-of-session notification with a summary when the stream processing session is complete.
Sourcepub fn kms_key_id(self, input: impl Into<String>) -> Self
pub fn kms_key_id(self, input: impl Into<String>) -> Self
The identifier for your AWS Key Management Service key (AWS KMS key). This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor. You can supply the Amazon Resource Name (ARN) of your KMS key, the ID of your KMS key, an alias for your KMS key, or an alias ARN. The key is used to encrypt results and data published to your Amazon S3 bucket, which includes image frames and hero images. Your source images are unaffected.
Sourcepub fn set_kms_key_id(self, input: Option<String>) -> Self
pub fn set_kms_key_id(self, input: Option<String>) -> Self
The identifier for your AWS Key Management Service key (AWS KMS key). This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor. You can supply the Amazon Resource Name (ARN) of your KMS key, the ID of your KMS key, an alias for your KMS key, or an alias ARN. The key is used to encrypt results and data published to your Amazon S3 bucket, which includes image frames and hero images. Your source images are unaffected.
Sourcepub fn get_kms_key_id(&self) -> &Option<String>
pub fn get_kms_key_id(&self) -> &Option<String>
The identifier for your AWS Key Management Service key (AWS KMS key). This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor. You can supply the Amazon Resource Name (ARN) of your KMS key, the ID of your KMS key, an alias for your KMS key, or an alias ARN. The key is used to encrypt results and data published to your Amazon S3 bucket, which includes image frames and hero images. Your source images are unaffected.
Sourcepub fn regions_of_interest(self, input: RegionOfInterest) -> Self
pub fn regions_of_interest(self, input: RegionOfInterest) -> Self
Appends an item to RegionsOfInterest
.
To override the contents of this collection use set_regions_of_interest
.
Specifies locations in the frames where Amazon Rekognition checks for objects or people. You can specify up to 10 regions of interest, and each region has either a polygon or a bounding box. This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor.
Sourcepub fn set_regions_of_interest(
self,
input: Option<Vec<RegionOfInterest>>,
) -> Self
pub fn set_regions_of_interest( self, input: Option<Vec<RegionOfInterest>>, ) -> Self
Specifies locations in the frames where Amazon Rekognition checks for objects or people. You can specify up to 10 regions of interest, and each region has either a polygon or a bounding box. This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor.
Sourcepub fn get_regions_of_interest(&self) -> &Option<Vec<RegionOfInterest>>
pub fn get_regions_of_interest(&self) -> &Option<Vec<RegionOfInterest>>
Specifies locations in the frames where Amazon Rekognition checks for objects or people. You can specify up to 10 regions of interest, and each region has either a polygon or a bounding box. This is an optional parameter for label detection stream processors and should not be used to create a face search stream processor.
Sourcepub fn data_sharing_preference(
self,
input: StreamProcessorDataSharingPreference,
) -> Self
pub fn data_sharing_preference( self, input: StreamProcessorDataSharingPreference, ) -> Self
Shows whether you are sharing data with Rekognition to improve model performance. You can choose this option at the account level or on a per-stream basis. Note that if you opt out at the account level this setting is ignored on individual streams.
Sourcepub fn set_data_sharing_preference(
self,
input: Option<StreamProcessorDataSharingPreference>,
) -> Self
pub fn set_data_sharing_preference( self, input: Option<StreamProcessorDataSharingPreference>, ) -> Self
Shows whether you are sharing data with Rekognition to improve model performance. You can choose this option at the account level or on a per-stream basis. Note that if you opt out at the account level this setting is ignored on individual streams.
Sourcepub fn get_data_sharing_preference(
&self,
) -> &Option<StreamProcessorDataSharingPreference>
pub fn get_data_sharing_preference( &self, ) -> &Option<StreamProcessorDataSharingPreference>
Shows whether you are sharing data with Rekognition to improve model performance. You can choose this option at the account level or on a per-stream basis. Note that if you opt out at the account level this setting is ignored on individual streams.
Trait Implementations§
Source§impl Clone for CreateStreamProcessorFluentBuilder
impl Clone for CreateStreamProcessorFluentBuilder
Source§fn clone(&self) -> CreateStreamProcessorFluentBuilder
fn clone(&self) -> CreateStreamProcessorFluentBuilder
1.0.0 · Source§const fn clone_from(&mut self, source: &Self)
const fn clone_from(&mut self, source: &Self)
source
. Read moreAuto Trait Implementations§
impl Freeze for CreateStreamProcessorFluentBuilder
impl !RefUnwindSafe for CreateStreamProcessorFluentBuilder
impl Send for CreateStreamProcessorFluentBuilder
impl Sync for CreateStreamProcessorFluentBuilder
impl Unpin for CreateStreamProcessorFluentBuilder
impl !UnwindSafe for CreateStreamProcessorFluentBuilder
Blanket Implementations§
Source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
Source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Source§impl<T> CloneToUninit for Twhere
T: Clone,
impl<T> CloneToUninit for Twhere
T: Clone,
Source§impl<T> Instrument for T
impl<T> Instrument for T
Source§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
Source§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
Source§impl<T> IntoEither for T
impl<T> IntoEither for T
Source§fn into_either(self, into_left: bool) -> Either<Self, Self>
fn into_either(self, into_left: bool) -> Either<Self, Self>
self
into a Left
variant of Either<Self, Self>
if into_left
is true
.
Converts self
into a Right
variant of Either<Self, Self>
otherwise. Read moreSource§fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
self
into a Left
variant of Either<Self, Self>
if into_left(&self)
returns true
.
Converts self
into a Right
variant of Either<Self, Self>
otherwise. Read moreSource§impl<T> Paint for Twhere
T: ?Sized,
impl<T> Paint for Twhere
T: ?Sized,
Source§fn fg(&self, value: Color) -> Painted<&T>
fn fg(&self, value: Color) -> Painted<&T>
Returns a styled value derived from self
with the foreground set to
value
.
This method should be used rarely. Instead, prefer to use color-specific
builder methods like red()
and
green()
, which have the same functionality but are
pithier.
§Example
Set foreground color to white using fg()
:
use yansi::{Paint, Color};
painted.fg(Color::White);
Set foreground color to white using white()
.
use yansi::Paint;
painted.white();
Source§fn bright_black(&self) -> Painted<&T>
fn bright_black(&self) -> Painted<&T>
Source§fn bright_red(&self) -> Painted<&T>
fn bright_red(&self) -> Painted<&T>
Source§fn bright_green(&self) -> Painted<&T>
fn bright_green(&self) -> Painted<&T>
Source§fn bright_yellow(&self) -> Painted<&T>
fn bright_yellow(&self) -> Painted<&T>
Source§fn bright_blue(&self) -> Painted<&T>
fn bright_blue(&self) -> Painted<&T>
Source§fn bright_magenta(&self) -> Painted<&T>
fn bright_magenta(&self) -> Painted<&T>
Source§fn bright_cyan(&self) -> Painted<&T>
fn bright_cyan(&self) -> Painted<&T>
Source§fn bright_white(&self) -> Painted<&T>
fn bright_white(&self) -> Painted<&T>
Source§fn bg(&self, value: Color) -> Painted<&T>
fn bg(&self, value: Color) -> Painted<&T>
Returns a styled value derived from self
with the background set to
value
.
This method should be used rarely. Instead, prefer to use color-specific
builder methods like on_red()
and
on_green()
, which have the same functionality but
are pithier.
§Example
Set background color to red using fg()
:
use yansi::{Paint, Color};
painted.bg(Color::Red);
Set background color to red using on_red()
.
use yansi::Paint;
painted.on_red();
Source§fn on_primary(&self) -> Painted<&T>
fn on_primary(&self) -> Painted<&T>
Source§fn on_magenta(&self) -> Painted<&T>
fn on_magenta(&self) -> Painted<&T>
Source§fn on_bright_black(&self) -> Painted<&T>
fn on_bright_black(&self) -> Painted<&T>
Source§fn on_bright_red(&self) -> Painted<&T>
fn on_bright_red(&self) -> Painted<&T>
Source§fn on_bright_green(&self) -> Painted<&T>
fn on_bright_green(&self) -> Painted<&T>
Source§fn on_bright_yellow(&self) -> Painted<&T>
fn on_bright_yellow(&self) -> Painted<&T>
Source§fn on_bright_blue(&self) -> Painted<&T>
fn on_bright_blue(&self) -> Painted<&T>
Source§fn on_bright_magenta(&self) -> Painted<&T>
fn on_bright_magenta(&self) -> Painted<&T>
Source§fn on_bright_cyan(&self) -> Painted<&T>
fn on_bright_cyan(&self) -> Painted<&T>
Source§fn on_bright_white(&self) -> Painted<&T>
fn on_bright_white(&self) -> Painted<&T>
Source§fn attr(&self, value: Attribute) -> Painted<&T>
fn attr(&self, value: Attribute) -> Painted<&T>
Enables the styling Attribute
value
.
This method should be used rarely. Instead, prefer to use
attribute-specific builder methods like bold()
and
underline()
, which have the same functionality
but are pithier.
§Example
Make text bold using attr()
:
use yansi::{Paint, Attribute};
painted.attr(Attribute::Bold);
Make text bold using using bold()
.
use yansi::Paint;
painted.bold();
Source§fn rapid_blink(&self) -> Painted<&T>
fn rapid_blink(&self) -> Painted<&T>
Source§fn quirk(&self, value: Quirk) -> Painted<&T>
fn quirk(&self, value: Quirk) -> Painted<&T>
Enables the yansi
Quirk
value
.
This method should be used rarely. Instead, prefer to use quirk-specific
builder methods like mask()
and
wrap()
, which have the same functionality but are
pithier.
§Example
Enable wrapping using .quirk()
:
use yansi::{Paint, Quirk};
painted.quirk(Quirk::Wrap);
Enable wrapping using wrap()
.
use yansi::Paint;
painted.wrap();
Source§fn clear(&self) -> Painted<&T>
👎Deprecated since 1.0.1: renamed to resetting()
due to conflicts with Vec::clear()
.
The clear()
method will be removed in a future release.
fn clear(&self) -> Painted<&T>
resetting()
due to conflicts with Vec::clear()
.
The clear()
method will be removed in a future release.Source§fn whenever(&self, value: Condition) -> Painted<&T>
fn whenever(&self, value: Condition) -> Painted<&T>
Conditionally enable styling based on whether the Condition
value
applies. Replaces any previous condition.
See the crate level docs for more details.
§Example
Enable styling painted
only when both stdout
and stderr
are TTYs:
use yansi::{Paint, Condition};
painted.red().on_yellow().whenever(Condition::STDOUTERR_ARE_TTY);