#[non_exhaustive]pub struct GetModelInvocationJobOutputBuilder { /* private fields */ }
Expand description
A builder for GetModelInvocationJobOutput
.
Implementations§
Source§impl GetModelInvocationJobOutputBuilder
impl GetModelInvocationJobOutputBuilder
Sourcepub fn job_arn(self, input: impl Into<String>) -> Self
pub fn job_arn(self, input: impl Into<String>) -> Self
The Amazon Resource Name (ARN) of the batch inference job.
This field is required.Sourcepub fn set_job_arn(self, input: Option<String>) -> Self
pub fn set_job_arn(self, input: Option<String>) -> Self
The Amazon Resource Name (ARN) of the batch inference job.
Sourcepub fn get_job_arn(&self) -> &Option<String>
pub fn get_job_arn(&self) -> &Option<String>
The Amazon Resource Name (ARN) of the batch inference job.
Sourcepub fn set_job_name(self, input: Option<String>) -> Self
pub fn set_job_name(self, input: Option<String>) -> Self
The name of the batch inference job.
Sourcepub fn get_job_name(&self) -> &Option<String>
pub fn get_job_name(&self) -> &Option<String>
The name of the batch inference job.
Sourcepub fn model_id(self, input: impl Into<String>) -> Self
pub fn model_id(self, input: impl Into<String>) -> Self
The unique identifier of the foundation model used for model inference.
This field is required.Sourcepub fn set_model_id(self, input: Option<String>) -> Self
pub fn set_model_id(self, input: Option<String>) -> Self
The unique identifier of the foundation model used for model inference.
Sourcepub fn get_model_id(&self) -> &Option<String>
pub fn get_model_id(&self) -> &Option<String>
The unique identifier of the foundation model used for model inference.
Sourcepub fn client_request_token(self, input: impl Into<String>) -> Self
pub fn client_request_token(self, input: impl Into<String>) -> Self
A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
Sourcepub fn set_client_request_token(self, input: Option<String>) -> Self
pub fn set_client_request_token(self, input: Option<String>) -> Self
A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
Sourcepub fn get_client_request_token(&self) -> &Option<String>
pub fn get_client_request_token(&self) -> &Option<String>
A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
Sourcepub fn role_arn(self, input: impl Into<String>) -> Self
pub fn role_arn(self, input: impl Into<String>) -> Self
The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
This field is required.Sourcepub fn set_role_arn(self, input: Option<String>) -> Self
pub fn set_role_arn(self, input: Option<String>) -> Self
The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
Sourcepub fn get_role_arn(&self) -> &Option<String>
pub fn get_role_arn(&self) -> &Option<String>
The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
Sourcepub fn status(self, input: ModelInvocationJobStatus) -> Self
pub fn status(self, input: ModelInvocationJobStatus) -> Self
The status of the batch inference job.
The following statuses are possible:
-
Submitted – This job has been submitted to a queue for validation.
-
Validating – This job is being validated for the requirements described in Format and upload your batch inference data. The criteria include the following:
-
Your IAM service role has access to the Amazon S3 buckets containing your files.
-
Your files are .jsonl files and each individual record is a JSON object in the correct format. Note that validation doesn't check if the
modelInput
value matches the request body for the model. -
Your files fulfill the requirements for file size and number of records. For more information, see Quotas for Amazon Bedrock.
-
-
Scheduled – This job has been validated and is now in a queue. The job will automatically start when it reaches its turn.
-
Expired – This job timed out because it was scheduled but didn't begin before the set timeout duration. Submit a new job request.
-
InProgress – This job has begun. You can start viewing the results in the output S3 location.
-
Completed – This job has successfully completed. View the output files in the output S3 location.
-
PartiallyCompleted – This job has partially completed. Not all of your records could be processed in time. View the output files in the output S3 location.
-
Failed – This job has failed. Check the failure message for any further details. For further assistance, reach out to the Amazon Web ServicesSupport Center.
-
Stopped – This job was stopped by a user.
-
Stopping – This job is being stopped by a user.
Sourcepub fn set_status(self, input: Option<ModelInvocationJobStatus>) -> Self
pub fn set_status(self, input: Option<ModelInvocationJobStatus>) -> Self
The status of the batch inference job.
The following statuses are possible:
-
Submitted – This job has been submitted to a queue for validation.
-
Validating – This job is being validated for the requirements described in Format and upload your batch inference data. The criteria include the following:
-
Your IAM service role has access to the Amazon S3 buckets containing your files.
-
Your files are .jsonl files and each individual record is a JSON object in the correct format. Note that validation doesn't check if the
modelInput
value matches the request body for the model. -
Your files fulfill the requirements for file size and number of records. For more information, see Quotas for Amazon Bedrock.
-
-
Scheduled – This job has been validated and is now in a queue. The job will automatically start when it reaches its turn.
-
Expired – This job timed out because it was scheduled but didn't begin before the set timeout duration. Submit a new job request.
-
InProgress – This job has begun. You can start viewing the results in the output S3 location.
-
Completed – This job has successfully completed. View the output files in the output S3 location.
-
PartiallyCompleted – This job has partially completed. Not all of your records could be processed in time. View the output files in the output S3 location.
-
Failed – This job has failed. Check the failure message for any further details. For further assistance, reach out to the Amazon Web ServicesSupport Center.
-
Stopped – This job was stopped by a user.
-
Stopping – This job is being stopped by a user.
Sourcepub fn get_status(&self) -> &Option<ModelInvocationJobStatus>
pub fn get_status(&self) -> &Option<ModelInvocationJobStatus>
The status of the batch inference job.
The following statuses are possible:
-
Submitted – This job has been submitted to a queue for validation.
-
Validating – This job is being validated for the requirements described in Format and upload your batch inference data. The criteria include the following:
-
Your IAM service role has access to the Amazon S3 buckets containing your files.
-
Your files are .jsonl files and each individual record is a JSON object in the correct format. Note that validation doesn't check if the
modelInput
value matches the request body for the model. -
Your files fulfill the requirements for file size and number of records. For more information, see Quotas for Amazon Bedrock.
-
-
Scheduled – This job has been validated and is now in a queue. The job will automatically start when it reaches its turn.
-
Expired – This job timed out because it was scheduled but didn't begin before the set timeout duration. Submit a new job request.
-
InProgress – This job has begun. You can start viewing the results in the output S3 location.
-
Completed – This job has successfully completed. View the output files in the output S3 location.
-
PartiallyCompleted – This job has partially completed. Not all of your records could be processed in time. View the output files in the output S3 location.
-
Failed – This job has failed. Check the failure message for any further details. For further assistance, reach out to the Amazon Web ServicesSupport Center.
-
Stopped – This job was stopped by a user.
-
Stopping – This job is being stopped by a user.
Sourcepub fn message(self, input: impl Into<String>) -> Self
pub fn message(self, input: impl Into<String>) -> Self
If the batch inference job failed, this field contains a message describing why the job failed.
Sourcepub fn set_message(self, input: Option<String>) -> Self
pub fn set_message(self, input: Option<String>) -> Self
If the batch inference job failed, this field contains a message describing why the job failed.
Sourcepub fn get_message(&self) -> &Option<String>
pub fn get_message(&self) -> &Option<String>
If the batch inference job failed, this field contains a message describing why the job failed.
Sourcepub fn submit_time(self, input: DateTime) -> Self
pub fn submit_time(self, input: DateTime) -> Self
The time at which the batch inference job was submitted.
This field is required.Sourcepub fn set_submit_time(self, input: Option<DateTime>) -> Self
pub fn set_submit_time(self, input: Option<DateTime>) -> Self
The time at which the batch inference job was submitted.
Sourcepub fn get_submit_time(&self) -> &Option<DateTime>
pub fn get_submit_time(&self) -> &Option<DateTime>
The time at which the batch inference job was submitted.
Sourcepub fn last_modified_time(self, input: DateTime) -> Self
pub fn last_modified_time(self, input: DateTime) -> Self
The time at which the batch inference job was last modified.
Sourcepub fn set_last_modified_time(self, input: Option<DateTime>) -> Self
pub fn set_last_modified_time(self, input: Option<DateTime>) -> Self
The time at which the batch inference job was last modified.
Sourcepub fn get_last_modified_time(&self) -> &Option<DateTime>
pub fn get_last_modified_time(&self) -> &Option<DateTime>
The time at which the batch inference job was last modified.
Sourcepub fn end_time(self, input: DateTime) -> Self
pub fn end_time(self, input: DateTime) -> Self
The time at which the batch inference job ended.
Sourcepub fn set_end_time(self, input: Option<DateTime>) -> Self
pub fn set_end_time(self, input: Option<DateTime>) -> Self
The time at which the batch inference job ended.
Sourcepub fn get_end_time(&self) -> &Option<DateTime>
pub fn get_end_time(&self) -> &Option<DateTime>
The time at which the batch inference job ended.
Sourcepub fn input_data_config(self, input: ModelInvocationJobInputDataConfig) -> Self
pub fn input_data_config(self, input: ModelInvocationJobInputDataConfig) -> Self
Details about the location of the input to the batch inference job.
This field is required.Sourcepub fn set_input_data_config(
self,
input: Option<ModelInvocationJobInputDataConfig>,
) -> Self
pub fn set_input_data_config( self, input: Option<ModelInvocationJobInputDataConfig>, ) -> Self
Details about the location of the input to the batch inference job.
Sourcepub fn get_input_data_config(
&self,
) -> &Option<ModelInvocationJobInputDataConfig>
pub fn get_input_data_config( &self, ) -> &Option<ModelInvocationJobInputDataConfig>
Details about the location of the input to the batch inference job.
Sourcepub fn output_data_config(
self,
input: ModelInvocationJobOutputDataConfig,
) -> Self
pub fn output_data_config( self, input: ModelInvocationJobOutputDataConfig, ) -> Self
Details about the location of the output of the batch inference job.
This field is required.Sourcepub fn set_output_data_config(
self,
input: Option<ModelInvocationJobOutputDataConfig>,
) -> Self
pub fn set_output_data_config( self, input: Option<ModelInvocationJobOutputDataConfig>, ) -> Self
Details about the location of the output of the batch inference job.
Sourcepub fn get_output_data_config(
&self,
) -> &Option<ModelInvocationJobOutputDataConfig>
pub fn get_output_data_config( &self, ) -> &Option<ModelInvocationJobOutputDataConfig>
Details about the location of the output of the batch inference job.
Sourcepub fn vpc_config(self, input: VpcConfig) -> Self
pub fn vpc_config(self, input: VpcConfig) -> Self
The configuration of the Virtual Private Cloud (VPC) for the data in the batch inference job. For more information, see Protect batch inference jobs using a VPC.
Sourcepub fn set_vpc_config(self, input: Option<VpcConfig>) -> Self
pub fn set_vpc_config(self, input: Option<VpcConfig>) -> Self
The configuration of the Virtual Private Cloud (VPC) for the data in the batch inference job. For more information, see Protect batch inference jobs using a VPC.
Sourcepub fn get_vpc_config(&self) -> &Option<VpcConfig>
pub fn get_vpc_config(&self) -> &Option<VpcConfig>
The configuration of the Virtual Private Cloud (VPC) for the data in the batch inference job. For more information, see Protect batch inference jobs using a VPC.
Sourcepub fn timeout_duration_in_hours(self, input: i32) -> Self
pub fn timeout_duration_in_hours(self, input: i32) -> Self
The number of hours after which batch inference job was set to time out.
Sourcepub fn set_timeout_duration_in_hours(self, input: Option<i32>) -> Self
pub fn set_timeout_duration_in_hours(self, input: Option<i32>) -> Self
The number of hours after which batch inference job was set to time out.
Sourcepub fn get_timeout_duration_in_hours(&self) -> &Option<i32>
pub fn get_timeout_duration_in_hours(&self) -> &Option<i32>
The number of hours after which batch inference job was set to time out.
Sourcepub fn job_expiration_time(self, input: DateTime) -> Self
pub fn job_expiration_time(self, input: DateTime) -> Self
The time at which the batch inference job times or timed out.
Sourcepub fn set_job_expiration_time(self, input: Option<DateTime>) -> Self
pub fn set_job_expiration_time(self, input: Option<DateTime>) -> Self
The time at which the batch inference job times or timed out.
Sourcepub fn get_job_expiration_time(&self) -> &Option<DateTime>
pub fn get_job_expiration_time(&self) -> &Option<DateTime>
The time at which the batch inference job times or timed out.
Sourcepub fn build(self) -> Result<GetModelInvocationJobOutput, BuildError>
pub fn build(self) -> Result<GetModelInvocationJobOutput, BuildError>
Consumes the builder and constructs a GetModelInvocationJobOutput
.
This method will fail if any of the following fields are not set:
Trait Implementations§
Source§impl Clone for GetModelInvocationJobOutputBuilder
impl Clone for GetModelInvocationJobOutputBuilder
Source§fn clone(&self) -> GetModelInvocationJobOutputBuilder
fn clone(&self) -> GetModelInvocationJobOutputBuilder
1.0.0 · Source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read moreSource§impl Default for GetModelInvocationJobOutputBuilder
impl Default for GetModelInvocationJobOutputBuilder
Source§fn default() -> GetModelInvocationJobOutputBuilder
fn default() -> GetModelInvocationJobOutputBuilder
Source§impl PartialEq for GetModelInvocationJobOutputBuilder
impl PartialEq for GetModelInvocationJobOutputBuilder
Source§fn eq(&self, other: &GetModelInvocationJobOutputBuilder) -> bool
fn eq(&self, other: &GetModelInvocationJobOutputBuilder) -> bool
self
and other
values to be equal, and is used by ==
.impl StructuralPartialEq for GetModelInvocationJobOutputBuilder
Auto Trait Implementations§
impl Freeze for GetModelInvocationJobOutputBuilder
impl RefUnwindSafe for GetModelInvocationJobOutputBuilder
impl Send for GetModelInvocationJobOutputBuilder
impl Sync for GetModelInvocationJobOutputBuilder
impl Unpin for GetModelInvocationJobOutputBuilder
impl UnwindSafe for GetModelInvocationJobOutputBuilder
Blanket Implementations§
Source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
Source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
Source§impl<T> CloneToUninit for Twhere
T: Clone,
impl<T> CloneToUninit for Twhere
T: Clone,
Source§impl<T> Instrument for T
impl<T> Instrument for T
Source§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
Source§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
Source§impl<T> IntoEither for T
impl<T> IntoEither for T
Source§fn into_either(self, into_left: bool) -> Either<Self, Self>
fn into_either(self, into_left: bool) -> Either<Self, Self>
self
into a Left
variant of Either<Self, Self>
if into_left
is true
.
Converts self
into a Right
variant of Either<Self, Self>
otherwise. Read moreSource§fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
self
into a Left
variant of Either<Self, Self>
if into_left(&self)
returns true
.
Converts self
into a Right
variant of Either<Self, Self>
otherwise. Read moreSource§impl<T> Paint for Twhere
T: ?Sized,
impl<T> Paint for Twhere
T: ?Sized,
Source§fn fg(&self, value: Color) -> Painted<&T>
fn fg(&self, value: Color) -> Painted<&T>
Returns a styled value derived from self
with the foreground set to
value
.
This method should be used rarely. Instead, prefer to use color-specific
builder methods like red()
and
green()
, which have the same functionality but are
pithier.
§Example
Set foreground color to white using fg()
:
use yansi::{Paint, Color};
painted.fg(Color::White);
Set foreground color to white using white()
.
use yansi::Paint;
painted.white();
Source§fn bright_black(&self) -> Painted<&T>
fn bright_black(&self) -> Painted<&T>
Source§fn bright_red(&self) -> Painted<&T>
fn bright_red(&self) -> Painted<&T>
Source§fn bright_green(&self) -> Painted<&T>
fn bright_green(&self) -> Painted<&T>
Source§fn bright_yellow(&self) -> Painted<&T>
fn bright_yellow(&self) -> Painted<&T>
Source§fn bright_blue(&self) -> Painted<&T>
fn bright_blue(&self) -> Painted<&T>
Source§fn bright_magenta(&self) -> Painted<&T>
fn bright_magenta(&self) -> Painted<&T>
Source§fn bright_cyan(&self) -> Painted<&T>
fn bright_cyan(&self) -> Painted<&T>
Source§fn bright_white(&self) -> Painted<&T>
fn bright_white(&self) -> Painted<&T>
Source§fn bg(&self, value: Color) -> Painted<&T>
fn bg(&self, value: Color) -> Painted<&T>
Returns a styled value derived from self
with the background set to
value
.
This method should be used rarely. Instead, prefer to use color-specific
builder methods like on_red()
and
on_green()
, which have the same functionality but
are pithier.
§Example
Set background color to red using fg()
:
use yansi::{Paint, Color};
painted.bg(Color::Red);
Set background color to red using on_red()
.
use yansi::Paint;
painted.on_red();
Source§fn on_primary(&self) -> Painted<&T>
fn on_primary(&self) -> Painted<&T>
Source§fn on_magenta(&self) -> Painted<&T>
fn on_magenta(&self) -> Painted<&T>
Source§fn on_bright_black(&self) -> Painted<&T>
fn on_bright_black(&self) -> Painted<&T>
Source§fn on_bright_red(&self) -> Painted<&T>
fn on_bright_red(&self) -> Painted<&T>
Source§fn on_bright_green(&self) -> Painted<&T>
fn on_bright_green(&self) -> Painted<&T>
Source§fn on_bright_yellow(&self) -> Painted<&T>
fn on_bright_yellow(&self) -> Painted<&T>
Source§fn on_bright_blue(&self) -> Painted<&T>
fn on_bright_blue(&self) -> Painted<&T>
Source§fn on_bright_magenta(&self) -> Painted<&T>
fn on_bright_magenta(&self) -> Painted<&T>
Source§fn on_bright_cyan(&self) -> Painted<&T>
fn on_bright_cyan(&self) -> Painted<&T>
Source§fn on_bright_white(&self) -> Painted<&T>
fn on_bright_white(&self) -> Painted<&T>
Source§fn attr(&self, value: Attribute) -> Painted<&T>
fn attr(&self, value: Attribute) -> Painted<&T>
Enables the styling Attribute
value
.
This method should be used rarely. Instead, prefer to use
attribute-specific builder methods like bold()
and
underline()
, which have the same functionality
but are pithier.
§Example
Make text bold using attr()
:
use yansi::{Paint, Attribute};
painted.attr(Attribute::Bold);
Make text bold using using bold()
.
use yansi::Paint;
painted.bold();
Source§fn rapid_blink(&self) -> Painted<&T>
fn rapid_blink(&self) -> Painted<&T>
Source§fn quirk(&self, value: Quirk) -> Painted<&T>
fn quirk(&self, value: Quirk) -> Painted<&T>
Enables the yansi
Quirk
value
.
This method should be used rarely. Instead, prefer to use quirk-specific
builder methods like mask()
and
wrap()
, which have the same functionality but are
pithier.
§Example
Enable wrapping using .quirk()
:
use yansi::{Paint, Quirk};
painted.quirk(Quirk::Wrap);
Enable wrapping using wrap()
.
use yansi::Paint;
painted.wrap();
Source§fn clear(&self) -> Painted<&T>
👎Deprecated since 1.0.1: renamed to resetting()
due to conflicts with Vec::clear()
.
The clear()
method will be removed in a future release.
fn clear(&self) -> Painted<&T>
resetting()
due to conflicts with Vec::clear()
.
The clear()
method will be removed in a future release.Source§fn whenever(&self, value: Condition) -> Painted<&T>
fn whenever(&self, value: Condition) -> Painted<&T>
Conditionally enable styling based on whether the Condition
value
applies. Replaces any previous condition.
See the crate level docs for more details.
§Example
Enable styling painted
only when both stdout
and stderr
are TTYs:
use yansi::{Paint, Condition};
painted.red().on_yellow().whenever(Condition::STDOUTERR_ARE_TTY);