Struct aws_sdk_sagemaker::model::Channel [−][src]
#[non_exhaustive]pub struct Channel {
pub channel_name: Option<String>,
pub data_source: Option<DataSource>,
pub content_type: Option<String>,
pub compression_type: Option<CompressionType>,
pub record_wrapper_type: Option<RecordWrapper>,
pub input_mode: Option<TrainingInputMode>,
pub shuffle_config: Option<ShuffleConfig>,
}
Expand description
A channel is a named input source that training algorithms can consume.
Fields (Non-exhaustive)
This struct is marked as non-exhaustive
Struct { .. }
syntax; cannot be matched against without a wildcard ..
; and struct update syntax will not work.channel_name: Option<String>
The name of the channel.
data_source: Option<DataSource>
The location of the channel data.
content_type: Option<String>
The MIME type of the data.
compression_type: Option<CompressionType>
If training data is compressed, the compression type. The default value is
None
. CompressionType
is used only in Pipe input mode. In
File mode, leave this field unset or set it to None.
record_wrapper_type: Option<RecordWrapper>
Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format. In this case, Amazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.
In File mode, leave this field unset or set it to None.
input_mode: Option<TrainingInputMode>
(Optional) The input mode to use for the data channel in a training job. If you don't
set a value for InputMode
, Amazon SageMaker uses the value set for
TrainingInputMode
. Use this parameter to override the
TrainingInputMode
setting in a AlgorithmSpecification
request when you have a channel that needs a different input mode from the training
job's general setting. To download the data from Amazon Simple Storage Service (Amazon S3) to the provisioned ML
storage volume, and mount the directory to a Docker volume, use File
input
mode. To stream data directly from Amazon S3 to the container, choose Pipe
input
mode.
To use a model for incremental training, choose File
input model.
shuffle_config: Option<ShuffleConfig>
A configuration for a shuffle option for input data in a channel. If you use
S3Prefix
for S3DataType
, this shuffles the results of the
S3 key prefix matches. If you use ManifestFile
, the order of the S3 object
references in the ManifestFile
is shuffled. If you use
AugmentedManifestFile
, the order of the JSON lines in the
AugmentedManifestFile
is shuffled. The shuffling order is determined
using the Seed
value.
For Pipe input mode, shuffling is done at the start of every epoch. With large
datasets this ensures that the order of the training data is different for each epoch,
it helps reduce bias and possible overfitting. In a multi-node training job when
ShuffleConfig is combined with S3DataDistributionType
of
ShardedByS3Key
, the data is shuffled across nodes so that the content
sent to a particular node on the first epoch might be sent to a different node on the
second epoch.
Implementations
The name of the channel.
The location of the channel data.
The MIME type of the data.
If training data is compressed, the compression type. The default value is
None
. CompressionType
is used only in Pipe input mode. In
File mode, leave this field unset or set it to None.
Specify RecordIO as the value when input data is in raw format but the training algorithm requires the RecordIO format. In this case, Amazon SageMaker wraps each individual S3 object in a RecordIO record. If the input data is already in RecordIO format, you don't need to set this attribute. For more information, see Create a Dataset Using RecordIO.
In File mode, leave this field unset or set it to None.
(Optional) The input mode to use for the data channel in a training job. If you don't
set a value for InputMode
, Amazon SageMaker uses the value set for
TrainingInputMode
. Use this parameter to override the
TrainingInputMode
setting in a AlgorithmSpecification
request when you have a channel that needs a different input mode from the training
job's general setting. To download the data from Amazon Simple Storage Service (Amazon S3) to the provisioned ML
storage volume, and mount the directory to a Docker volume, use File
input
mode. To stream data directly from Amazon S3 to the container, choose Pipe
input
mode.
To use a model for incremental training, choose File
input model.
A configuration for a shuffle option for input data in a channel. If you use
S3Prefix
for S3DataType
, this shuffles the results of the
S3 key prefix matches. If you use ManifestFile
, the order of the S3 object
references in the ManifestFile
is shuffled. If you use
AugmentedManifestFile
, the order of the JSON lines in the
AugmentedManifestFile
is shuffled. The shuffling order is determined
using the Seed
value.
For Pipe input mode, shuffling is done at the start of every epoch. With large
datasets this ensures that the order of the training data is different for each epoch,
it helps reduce bias and possible overfitting. In a multi-node training job when
ShuffleConfig is combined with S3DataDistributionType
of
ShardedByS3Key
, the data is shuffled across nodes so that the content
sent to a particular node on the first epoch might be sent to a different node on the
second epoch.
Trait Implementations
Auto Trait Implementations
impl RefUnwindSafe for Channel
impl UnwindSafe for Channel
Blanket Implementations
Mutably borrows from an owned value. Read more
Attaches the provided Subscriber
to this type, returning a
WithDispatch
wrapper. Read more
Attaches the current default Subscriber
to this type, returning a
WithDispatch
wrapper. Read more