Struct aws_sdk_sagemaker::types::TextGenerationJobConfig
source · #[non_exhaustive]pub struct TextGenerationJobConfig {
pub completion_criteria: Option<AutoMlJobCompletionCriteria>,
pub base_model_name: Option<String>,
pub text_generation_hyper_parameters: Option<HashMap<String, String>>,
}
Expand description
The collection of settings used by an AutoML job V2 for the text generation problem type.
The text generation models that support fine-tuning in Autopilot are currently accessible exclusively in regions supported by Canvas. Refer to the documentation of Canvas for the full list of its supported Regions.
Fields (Non-exhaustive)§
This struct is marked as non-exhaustive
Struct { .. }
syntax; cannot be matched against without a wildcard ..
; and struct update syntax will not work.completion_criteria: Option<AutoMlJobCompletionCriteria>
How long a fine-tuning job is allowed to run. For TextGenerationJobConfig
problem types, the MaxRuntimePerTrainingJobInSeconds
attribute of AutoMLJobCompletionCriteria
defaults to 72h (259200s).
base_model_name: Option<String>
The name of the base model to fine-tune. Autopilot supports fine-tuning a variety of large language models. For information on the list of supported models, see Text generation models supporting fine-tuning in Autopilot. If no BaseModelName
is provided, the default model used is Falcon7BInstruct.
text_generation_hyper_parameters: Option<HashMap<String, String>>
The hyperparameters used to configure and optimize the learning process of the base model. You can set any combination of the following hyperparameters for all base models. For more information on each supported hyperparameter, see Optimize the learning process of your text generation models with hyperparameters.
-
"epochCount"
: The number of times the model goes through the entire training dataset. Its value should be a string containing an integer value within the range of "1" to "10". -
"batchSize"
: The number of data samples used in each iteration of training. Its value should be a string containing an integer value within the range of "1" to "64". -
"learningRate"
: The step size at which a model's parameters are updated during training. Its value should be a string containing a floating-point value within the range of "0" to "1". -
"learningRateWarmupSteps"
: The number of training steps during which the learning rate gradually increases before reaching its target or maximum value. Its value should be a string containing an integer value within the range of "0" to "250".
Here is an example where all four hyperparameters are configured.
{ "epochCount":"5", "learningRate":"0.5", "batchSize": "32", "learningRateWarmupSteps": "10" }
Implementations§
source§impl TextGenerationJobConfig
impl TextGenerationJobConfig
sourcepub fn completion_criteria(&self) -> Option<&AutoMlJobCompletionCriteria>
pub fn completion_criteria(&self) -> Option<&AutoMlJobCompletionCriteria>
How long a fine-tuning job is allowed to run. For TextGenerationJobConfig
problem types, the MaxRuntimePerTrainingJobInSeconds
attribute of AutoMLJobCompletionCriteria
defaults to 72h (259200s).
sourcepub fn base_model_name(&self) -> Option<&str>
pub fn base_model_name(&self) -> Option<&str>
The name of the base model to fine-tune. Autopilot supports fine-tuning a variety of large language models. For information on the list of supported models, see Text generation models supporting fine-tuning in Autopilot. If no BaseModelName
is provided, the default model used is Falcon7BInstruct.
sourcepub fn text_generation_hyper_parameters(
&self
) -> Option<&HashMap<String, String>>
pub fn text_generation_hyper_parameters( &self ) -> Option<&HashMap<String, String>>
The hyperparameters used to configure and optimize the learning process of the base model. You can set any combination of the following hyperparameters for all base models. For more information on each supported hyperparameter, see Optimize the learning process of your text generation models with hyperparameters.
-
"epochCount"
: The number of times the model goes through the entire training dataset. Its value should be a string containing an integer value within the range of "1" to "10". -
"batchSize"
: The number of data samples used in each iteration of training. Its value should be a string containing an integer value within the range of "1" to "64". -
"learningRate"
: The step size at which a model's parameters are updated during training. Its value should be a string containing a floating-point value within the range of "0" to "1". -
"learningRateWarmupSteps"
: The number of training steps during which the learning rate gradually increases before reaching its target or maximum value. Its value should be a string containing an integer value within the range of "0" to "250".
Here is an example where all four hyperparameters are configured.
{ "epochCount":"5", "learningRate":"0.5", "batchSize": "32", "learningRateWarmupSteps": "10" }
source§impl TextGenerationJobConfig
impl TextGenerationJobConfig
sourcepub fn builder() -> TextGenerationJobConfigBuilder
pub fn builder() -> TextGenerationJobConfigBuilder
Creates a new builder-style object to manufacture TextGenerationJobConfig
.
Trait Implementations§
source§impl Clone for TextGenerationJobConfig
impl Clone for TextGenerationJobConfig
source§fn clone(&self) -> TextGenerationJobConfig
fn clone(&self) -> TextGenerationJobConfig
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read moresource§impl Debug for TextGenerationJobConfig
impl Debug for TextGenerationJobConfig
source§impl PartialEq for TextGenerationJobConfig
impl PartialEq for TextGenerationJobConfig
source§fn eq(&self, other: &TextGenerationJobConfig) -> bool
fn eq(&self, other: &TextGenerationJobConfig) -> bool
self
and other
values to be equal, and is used
by ==
.