pub struct CreateModelFluentBuilder { /* private fields */ }
Expand description
Fluent builder constructing a request to CreateModel
.
Creates a model in SageMaker. In the request, you name the model and describe a primary container. For the primary container, you specify the Docker image that contains inference code, artifacts (from prior training), and a custom environment map that the inference code uses when you deploy the model for predictions.
Use this API to create a model if you want to use SageMaker hosting services or run a batch transform job.
To host your model, you create an endpoint configuration with the CreateEndpointConfig
API, and then create an endpoint with the CreateEndpoint
API. SageMaker then deploys all of the containers that you defined for the model in the hosting environment.
For an example that calls this method when deploying a model to SageMaker hosting services, see Create a Model (Amazon Web Services SDK for Python (Boto 3)).
To run a batch transform using your model, you start a job with the CreateTransformJob
API. SageMaker uses your model and your dataset to get inferences which are then saved to a specified S3 location.
In the request, you also provide an IAM role that SageMaker can assume to access model artifacts and docker image for deployment on ML compute hosting instances or for batch transform jobs. In addition, you also use the IAM role to manage permissions the inference code needs. For example, if the inference code access any other Amazon Web Services resources, you grant necessary permissions via this role.
Implementations§
source§impl CreateModelFluentBuilder
impl CreateModelFluentBuilder
sourcepub fn as_input(&self) -> &CreateModelInputBuilder
pub fn as_input(&self) -> &CreateModelInputBuilder
Access the CreateModel as a reference.
sourcepub async fn send(
self
) -> Result<CreateModelOutput, SdkError<CreateModelError, HttpResponse>>
pub async fn send( self ) -> Result<CreateModelOutput, SdkError<CreateModelError, HttpResponse>>
Sends the request and returns the response.
If an error occurs, an SdkError
will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
sourcepub fn customize(
self
) -> CustomizableOperation<CreateModelOutput, CreateModelError, Self>
pub fn customize( self ) -> CustomizableOperation<CreateModelOutput, CreateModelError, Self>
Consumes this builder, creating a customizable operation that can be modified before being sent.
sourcepub fn model_name(self, input: impl Into<String>) -> Self
pub fn model_name(self, input: impl Into<String>) -> Self
The name of the new model.
sourcepub fn set_model_name(self, input: Option<String>) -> Self
pub fn set_model_name(self, input: Option<String>) -> Self
The name of the new model.
sourcepub fn get_model_name(&self) -> &Option<String>
pub fn get_model_name(&self) -> &Option<String>
The name of the new model.
sourcepub fn primary_container(self, input: ContainerDefinition) -> Self
pub fn primary_container(self, input: ContainerDefinition) -> Self
The location of the primary docker image containing inference code, associated artifacts, and custom environment map that the inference code uses when the model is deployed for predictions.
sourcepub fn set_primary_container(self, input: Option<ContainerDefinition>) -> Self
pub fn set_primary_container(self, input: Option<ContainerDefinition>) -> Self
The location of the primary docker image containing inference code, associated artifacts, and custom environment map that the inference code uses when the model is deployed for predictions.
sourcepub fn get_primary_container(&self) -> &Option<ContainerDefinition>
pub fn get_primary_container(&self) -> &Option<ContainerDefinition>
The location of the primary docker image containing inference code, associated artifacts, and custom environment map that the inference code uses when the model is deployed for predictions.
sourcepub fn containers(self, input: ContainerDefinition) -> Self
pub fn containers(self, input: ContainerDefinition) -> Self
Appends an item to Containers
.
To override the contents of this collection use set_containers
.
Specifies the containers in the inference pipeline.
sourcepub fn set_containers(self, input: Option<Vec<ContainerDefinition>>) -> Self
pub fn set_containers(self, input: Option<Vec<ContainerDefinition>>) -> Self
Specifies the containers in the inference pipeline.
sourcepub fn get_containers(&self) -> &Option<Vec<ContainerDefinition>>
pub fn get_containers(&self) -> &Option<Vec<ContainerDefinition>>
Specifies the containers in the inference pipeline.
sourcepub fn inference_execution_config(self, input: InferenceExecutionConfig) -> Self
pub fn inference_execution_config(self, input: InferenceExecutionConfig) -> Self
Specifies details of how containers in a multi-container endpoint are called.
sourcepub fn set_inference_execution_config(
self,
input: Option<InferenceExecutionConfig>
) -> Self
pub fn set_inference_execution_config( self, input: Option<InferenceExecutionConfig> ) -> Self
Specifies details of how containers in a multi-container endpoint are called.
sourcepub fn get_inference_execution_config(
&self
) -> &Option<InferenceExecutionConfig>
pub fn get_inference_execution_config( &self ) -> &Option<InferenceExecutionConfig>
Specifies details of how containers in a multi-container endpoint are called.
sourcepub fn execution_role_arn(self, input: impl Into<String>) -> Self
pub fn execution_role_arn(self, input: impl Into<String>) -> Self
The Amazon Resource Name (ARN) of the IAM role that SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances or for batch transform jobs. Deploying on ML compute instances is part of model hosting. For more information, see SageMaker Roles.
To be able to pass this role to SageMaker, the caller of this API must have the iam:PassRole
permission.
sourcepub fn set_execution_role_arn(self, input: Option<String>) -> Self
pub fn set_execution_role_arn(self, input: Option<String>) -> Self
The Amazon Resource Name (ARN) of the IAM role that SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances or for batch transform jobs. Deploying on ML compute instances is part of model hosting. For more information, see SageMaker Roles.
To be able to pass this role to SageMaker, the caller of this API must have the iam:PassRole
permission.
sourcepub fn get_execution_role_arn(&self) -> &Option<String>
pub fn get_execution_role_arn(&self) -> &Option<String>
The Amazon Resource Name (ARN) of the IAM role that SageMaker can assume to access model artifacts and docker image for deployment on ML compute instances or for batch transform jobs. Deploying on ML compute instances is part of model hosting. For more information, see SageMaker Roles.
To be able to pass this role to SageMaker, the caller of this API must have the iam:PassRole
permission.
Appends an item to Tags
.
To override the contents of this collection use set_tags
.
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
An array of key-value pairs. You can use tags to categorize your Amazon Web Services resources in different ways, for example, by purpose, owner, or environment. For more information, see Tagging Amazon Web Services Resources.
sourcepub fn vpc_config(self, input: VpcConfig) -> Self
pub fn vpc_config(self, input: VpcConfig) -> Self
A VpcConfig object that specifies the VPC that you want your model to connect to. Control access to and from your model container by configuring the VPC. VpcConfig
is used in hosting services and in batch transform. For more information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch Transform Jobs by Using an Amazon Virtual Private Cloud.
sourcepub fn set_vpc_config(self, input: Option<VpcConfig>) -> Self
pub fn set_vpc_config(self, input: Option<VpcConfig>) -> Self
A VpcConfig object that specifies the VPC that you want your model to connect to. Control access to and from your model container by configuring the VPC. VpcConfig
is used in hosting services and in batch transform. For more information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch Transform Jobs by Using an Amazon Virtual Private Cloud.
sourcepub fn get_vpc_config(&self) -> &Option<VpcConfig>
pub fn get_vpc_config(&self) -> &Option<VpcConfig>
A VpcConfig object that specifies the VPC that you want your model to connect to. Control access to and from your model container by configuring the VPC. VpcConfig
is used in hosting services and in batch transform. For more information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch Transform Jobs by Using an Amazon Virtual Private Cloud.
sourcepub fn enable_network_isolation(self, input: bool) -> Self
pub fn enable_network_isolation(self, input: bool) -> Self
Isolates the model container. No inbound or outbound network calls can be made to or from the model container.
sourcepub fn set_enable_network_isolation(self, input: Option<bool>) -> Self
pub fn set_enable_network_isolation(self, input: Option<bool>) -> Self
Isolates the model container. No inbound or outbound network calls can be made to or from the model container.
sourcepub fn get_enable_network_isolation(&self) -> &Option<bool>
pub fn get_enable_network_isolation(&self) -> &Option<bool>
Isolates the model container. No inbound or outbound network calls can be made to or from the model container.
Trait Implementations§
source§impl Clone for CreateModelFluentBuilder
impl Clone for CreateModelFluentBuilder
source§fn clone(&self) -> CreateModelFluentBuilder
fn clone(&self) -> CreateModelFluentBuilder
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read more