aws_sdk_cleanroomsml/client/get_trained_model_inference_job.rs
1// Code generated by software.amazon.smithy.rust.codegen.smithy-rs. DO NOT EDIT.
2impl super::Client {
3 /// Constructs a fluent builder for the [`GetTrainedModelInferenceJob`](crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder) operation.
4 ///
5 /// - The fluent builder is configurable:
6 /// - [`membership_identifier(impl Into<String>)`](crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder::membership_identifier) / [`set_membership_identifier(Option<String>)`](crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder::set_membership_identifier):<br>required: **true**<br><p>Provides the membership ID of the membership that contains the trained model inference job that you are interested in.</p><br>
7 /// - [`trained_model_inference_job_arn(impl Into<String>)`](crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder::trained_model_inference_job_arn) / [`set_trained_model_inference_job_arn(Option<String>)`](crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder::set_trained_model_inference_job_arn):<br>required: **true**<br><p>Provides the Amazon Resource Name (ARN) of the trained model inference job that you are interested in.</p><br>
8 /// - On success, responds with [`GetTrainedModelInferenceJobOutput`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput) with field(s):
9 /// - [`create_time(DateTime)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::create_time): <p>The time at which the trained model inference job was created.</p>
10 /// - [`update_time(DateTime)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::update_time): <p>The most recent time at which the trained model inference job was updated.</p>
11 /// - [`trained_model_inference_job_arn(String)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::trained_model_inference_job_arn): <p>The Amazon Resource Name (ARN) of the trained model inference job.</p>
12 /// - [`configured_model_algorithm_association_arn(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::configured_model_algorithm_association_arn): <p>The Amazon Resource Name (ARN) of the configured model algorithm association that was used for the trained model inference job.</p>
13 /// - [`name(String)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::name): <p>The name of the trained model inference job.</p>
14 /// - [`status(TrainedModelInferenceJobStatus)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::status): <p>The status of the trained model inference job.</p>
15 /// - [`trained_model_arn(String)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::trained_model_arn): <p>The Amazon Resource Name (ARN) for the trained model that was used for the trained model inference job.</p>
16 /// - [`trained_model_version_identifier(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::trained_model_version_identifier): <p>The version identifier of the trained model used for this inference job. This identifies the specific version of the trained model that was used to generate the inference results.</p>
17 /// - [`resource_config(Option<InferenceResourceConfig>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::resource_config): <p>The resource configuration information for the trained model inference job.</p>
18 /// - [`output_configuration(Option<InferenceOutputConfiguration>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::output_configuration): <p>The output configuration information for the trained model inference job.</p>
19 /// - [`membership_identifier(String)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::membership_identifier): <p>The membership ID of the membership that contains the trained model inference job.</p>
20 /// - [`data_source(Option<ModelInferenceDataSource>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::data_source): <p>The data source that was used for the trained model inference job.</p>
21 /// - [`container_execution_parameters(Option<InferenceContainerExecutionParameters>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::container_execution_parameters): <p>The execution parameters for the model inference job container.</p>
22 /// - [`status_details(Option<StatusDetails>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::status_details): <p>Details about the status of a resource.</p>
23 /// - [`description(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::description): <p>The description of the trained model inference job.</p>
24 /// - [`inference_container_image_digest(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::inference_container_image_digest): <p>Information about the training container image.</p>
25 /// - [`environment(Option<HashMap::<String, String>>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::environment): <p>The environment variables to set in the Docker container.</p>
26 /// - [`kms_key_arn(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::kms_key_arn): <p>The Amazon Resource Name (ARN) of the KMS key. This key is used to encrypt and decrypt customer-owned data in the ML inference job and associated data.</p>
27 /// - [`metrics_status(Option<MetricsStatus>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::metrics_status): <p>The metrics status for the trained model inference job.</p>
28 /// - [`metrics_status_details(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::metrics_status_details): <p>Details about the metrics status for the trained model inference job.</p>
29 /// - [`logs_status(Option<LogsStatus>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::logs_status): <p>The logs status for the trained model inference job.</p>
30 /// - [`logs_status_details(Option<String>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::logs_status_details): <p>Details about the logs status for the trained model inference job.</p>
31 /// - [`tags(Option<HashMap::<String, String>>)`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobOutput::tags): <p>The optional metadata that you applied to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define.</p> <p>The following basic restrictions apply to tags:</p> <ul> <li> <p>Maximum number of tags per resource - 50.</p></li> <li> <p>For each resource, each tag key must be unique, and each tag key can have only one value.</p></li> <li> <p>Maximum key length - 128 Unicode characters in UTF-8.</p></li> <li> <p>Maximum value length - 256 Unicode characters in UTF-8.</p></li> <li> <p>If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.</p></li> <li> <p>Tag keys and values are case sensitive.</p></li> <li> <p>Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms ML considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.</p></li> </ul>
32 /// - On failure, responds with [`SdkError<GetTrainedModelInferenceJobError>`](crate::operation::get_trained_model_inference_job::GetTrainedModelInferenceJobError)
33 pub fn get_trained_model_inference_job(
34 &self,
35 ) -> crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder {
36 crate::operation::get_trained_model_inference_job::builders::GetTrainedModelInferenceJobFluentBuilder::new(self.handle.clone())
37 }
38}