Struct aws_sdk_sagemaker::model::ProductionVariant [−][src]
#[non_exhaustive]pub struct ProductionVariant {
pub variant_name: Option<String>,
pub model_name: Option<String>,
pub initial_instance_count: Option<i32>,
pub instance_type: Option<ProductionVariantInstanceType>,
pub initial_variant_weight: Option<f32>,
pub accelerator_type: Option<ProductionVariantAcceleratorType>,
pub core_dump_config: Option<ProductionVariantCoreDumpConfig>,
}
Expand description
Identifies a model that you want to host and the resources chosen to deploy for hosting it. If you are deploying multiple models, tell Amazon SageMaker how to distribute traffic among the models by specifying variant weights.
Fields (Non-exhaustive)
This struct is marked as non-exhaustive
Struct { .. }
syntax; cannot be matched against without a wildcard ..
; and struct update syntax will not work.variant_name: Option<String>
The name of the production variant.
model_name: Option<String>
The name of the model that you want to host. This is the name that you specified when creating the model.
initial_instance_count: Option<i32>
Number of instances to launch initially.
instance_type: Option<ProductionVariantInstanceType>
The ML compute instance type.
initial_variant_weight: Option<f32>
Determines initial traffic distribution among all of the models that you specify in
the endpoint configuration. The traffic to a production variant is determined by the
ratio of the VariantWeight
to the sum of all VariantWeight
values across all ProductionVariants. If unspecified, it defaults to 1.0.
accelerator_type: Option<ProductionVariantAcceleratorType>
The size of the Elastic Inference (EI) instance to use for the production variant. EI instances provide on-demand GPU computing for inference. For more information, see Using Elastic Inference in Amazon SageMaker.
core_dump_config: Option<ProductionVariantCoreDumpConfig>
Specifies configuration for a core dump from the model container when the process crashes.
Implementations
The name of the production variant.
The name of the model that you want to host. This is the name that you specified when creating the model.
Number of instances to launch initially.
The ML compute instance type.
Determines initial traffic distribution among all of the models that you specify in
the endpoint configuration. The traffic to a production variant is determined by the
ratio of the VariantWeight
to the sum of all VariantWeight
values across all ProductionVariants. If unspecified, it defaults to 1.0.
The size of the Elastic Inference (EI) instance to use for the production variant. EI instances provide on-demand GPU computing for inference. For more information, see Using Elastic Inference in Amazon SageMaker.
Specifies configuration for a core dump from the model container when the process crashes.
Creates a new builder-style object to manufacture ProductionVariant
Trait Implementations
This method tests for self
and other
values to be equal, and is used
by ==
. Read more
This method tests for !=
.
Auto Trait Implementations
impl RefUnwindSafe for ProductionVariant
impl Send for ProductionVariant
impl Sync for ProductionVariant
impl Unpin for ProductionVariant
impl UnwindSafe for ProductionVariant
Blanket Implementations
Mutably borrows from an owned value. Read more
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more