Struct aws_sdk_sagemaker::model::TrainingJobDefinition
source · [−]#[non_exhaustive]pub struct TrainingJobDefinition {
pub training_input_mode: Option<TrainingInputMode>,
pub hyper_parameters: Option<HashMap<String, String>>,
pub input_data_config: Option<Vec<Channel>>,
pub output_data_config: Option<OutputDataConfig>,
pub resource_config: Option<ResourceConfig>,
pub stopping_condition: Option<StoppingCondition>,
}
Expand description
Defines the input needed to run a training job using the algorithm.
Fields (Non-exhaustive)
This struct is marked as non-exhaustive
Struct { .. }
syntax; cannot be matched against without a wildcard ..
; and struct update syntax will not work.training_input_mode: Option<TrainingInputMode>
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe
mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File
mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile
mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile
mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
hyper_parameters: Option<HashMap<String, String>>
The hyperparameters used for the training job.
input_data_config: Option<Vec<Channel>>
An array of Channel
objects, each of which specifies an input source.
output_data_config: Option<OutputDataConfig>
the path to the S3 bucket where you want to store model artifacts. Amazon SageMaker creates subfolders for the artifacts.
resource_config: Option<ResourceConfig>
The resources, including the ML compute instances and ML storage volumes, to use for model training.
stopping_condition: Option<StoppingCondition>
Specifies a limit to how long a model training job can run. It also specifies how long a managed Spot training job has to complete. When the job reaches the time limit, Amazon SageMaker ends the training job. Use this API to cap model training costs.
To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms can use this 120-second window to save the model artifacts.
Implementations
sourceimpl TrainingJobDefinition
impl TrainingJobDefinition
sourcepub fn training_input_mode(&self) -> Option<&TrainingInputMode>
pub fn training_input_mode(&self) -> Option<&TrainingInputMode>
The training input mode that the algorithm supports. For more information about input modes, see Algorithms.
Pipe mode
If an algorithm supports Pipe
mode, Amazon SageMaker streams data directly from Amazon S3 to the container.
File mode
If an algorithm supports File
mode, SageMaker downloads the training data from S3 to the provisioned ML storage volume, and mounts the directory to the Docker volume for the training container.
You must provision the ML storage volume with sufficient capacity to accommodate the data downloaded from S3. In addition to the training data, the ML storage volume also stores the output model. The algorithm container uses the ML storage volume to also store intermediate information, if any.
For distributed algorithms, training data is distributed uniformly. Your training duration is predictable if the input data objects sizes are approximately the same. SageMaker does not split the files any further for model training. If the object sizes are skewed, training won't be optimal as the data distribution is also skewed when one host in a training cluster is overloaded, thus becoming a bottleneck in training.
FastFile mode
If an algorithm supports FastFile
mode, SageMaker streams data directly from S3 to the container with no code changes, and provides file system access to the data. Users can author their training script to interact with these files as if they were stored on disk.
FastFile
mode works best when the data is read sequentially. Augmented manifest files aren't supported. The startup time is lower when there are fewer files in the S3 bucket provided.
sourcepub fn hyper_parameters(&self) -> Option<&HashMap<String, String>>
pub fn hyper_parameters(&self) -> Option<&HashMap<String, String>>
The hyperparameters used for the training job.
sourcepub fn input_data_config(&self) -> Option<&[Channel]>
pub fn input_data_config(&self) -> Option<&[Channel]>
An array of Channel
objects, each of which specifies an input source.
sourcepub fn output_data_config(&self) -> Option<&OutputDataConfig>
pub fn output_data_config(&self) -> Option<&OutputDataConfig>
the path to the S3 bucket where you want to store model artifacts. Amazon SageMaker creates subfolders for the artifacts.
sourcepub fn resource_config(&self) -> Option<&ResourceConfig>
pub fn resource_config(&self) -> Option<&ResourceConfig>
The resources, including the ML compute instances and ML storage volumes, to use for model training.
sourcepub fn stopping_condition(&self) -> Option<&StoppingCondition>
pub fn stopping_condition(&self) -> Option<&StoppingCondition>
Specifies a limit to how long a model training job can run. It also specifies how long a managed Spot training job has to complete. When the job reaches the time limit, Amazon SageMaker ends the training job. Use this API to cap model training costs.
To stop a job, Amazon SageMaker sends the algorithm the SIGTERM signal, which delays job termination for 120 seconds. Algorithms can use this 120-second window to save the model artifacts.
sourceimpl TrainingJobDefinition
impl TrainingJobDefinition
sourcepub fn builder() -> Builder
pub fn builder() -> Builder
Creates a new builder-style object to manufacture TrainingJobDefinition
Trait Implementations
sourceimpl Clone for TrainingJobDefinition
impl Clone for TrainingJobDefinition
sourcefn clone(&self) -> TrainingJobDefinition
fn clone(&self) -> TrainingJobDefinition
Returns a copy of the value. Read more
1.0.0 · sourcefn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
Performs copy-assignment from source
. Read more
sourceimpl Debug for TrainingJobDefinition
impl Debug for TrainingJobDefinition
sourceimpl PartialEq<TrainingJobDefinition> for TrainingJobDefinition
impl PartialEq<TrainingJobDefinition> for TrainingJobDefinition
sourcefn eq(&self, other: &TrainingJobDefinition) -> bool
fn eq(&self, other: &TrainingJobDefinition) -> bool
This method tests for self
and other
values to be equal, and is used
by ==
. Read more
sourcefn ne(&self, other: &TrainingJobDefinition) -> bool
fn ne(&self, other: &TrainingJobDefinition) -> bool
This method tests for !=
.
impl StructuralPartialEq for TrainingJobDefinition
Auto Trait Implementations
impl RefUnwindSafe for TrainingJobDefinition
impl Send for TrainingJobDefinition
impl Sync for TrainingJobDefinition
impl Unpin for TrainingJobDefinition
impl UnwindSafe for TrainingJobDefinition
Blanket Implementations
sourceimpl<T> BorrowMut<T> for T where
T: ?Sized,
impl<T> BorrowMut<T> for T where
T: ?Sized,
const: unstable · sourcepub fn borrow_mut(&mut self) -> &mut T
pub fn borrow_mut(&mut self) -> &mut T
Mutably borrows from an owned value. Read more
sourceimpl<T> Instrument for T
impl<T> Instrument for T
sourcefn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
sourcefn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
sourceimpl<T> ToOwned for T where
T: Clone,
impl<T> ToOwned for T where
T: Clone,
type Owned = T
type Owned = T
The resulting type after obtaining ownership.
sourcepub fn to_owned(&self) -> T
pub fn to_owned(&self) -> T
Creates owned data from borrowed data, usually by cloning. Read more
sourcepub fn clone_into(&self, target: &mut T)
pub fn clone_into(&self, target: &mut T)
toowned_clone_into
)Uses borrowed data to replace owned data, usually by cloning. Read more
sourceimpl<T> WithSubscriber for T
impl<T> WithSubscriber for T
sourcefn with_subscriber<S>(self, subscriber: S) -> WithDispatch<Self> where
S: Into<Dispatch>,
fn with_subscriber<S>(self, subscriber: S) -> WithDispatch<Self> where
S: Into<Dispatch>,
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
sourcefn with_current_subscriber(self) -> WithDispatch<Self>
fn with_current_subscriber(self) -> WithDispatch<Self>
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more