Struct aws_sdk_forecast::client::fluent_builders::CreateDataset
source · [−]pub struct CreateDataset { /* private fields */ }Expand description
Fluent builder constructing a request to CreateDataset.
Creates an Amazon Forecast dataset. The information about the dataset that you provide helps Forecast understand how to consume the data for model training. This includes the following:
-
DataFrequency- How frequently your historical time-series data is collected. -
DomainandDatasetType- Each dataset has an associated dataset domain and a type within the domain. Amazon Forecast provides a list of predefined domains and types within each domain. For each unique dataset domain and type within the domain, Amazon Forecast requires your data to include a minimum set of predefined fields. -
Schema- A schema specifies the fields in the dataset, including the field name and data type.
After creating a dataset, you import your training data into it and add the dataset to a dataset group. You use the dataset group to create a predictor. For more information, see Importing datasets.
To get a list of all your datasets, use the ListDatasets operation.
For example Forecast datasets, see the Amazon Forecast Sample GitHub repository.
The Status of a dataset must be ACTIVE before you can import training data. Use the DescribeDataset operation to get the status.
Implementations
sourceimpl CreateDataset
impl CreateDataset
sourcepub async fn customize(
self
) -> Result<CustomizableOperation<CreateDataset, AwsResponseRetryClassifier>, SdkError<CreateDatasetError>>
pub async fn customize(
self
) -> Result<CustomizableOperation<CreateDataset, AwsResponseRetryClassifier>, SdkError<CreateDatasetError>>
Consume this builder, creating a customizable operation that can be modified before being sent. The operation’s inner http::Request can be modified as well.
sourcepub async fn send(
self
) -> Result<CreateDatasetOutput, SdkError<CreateDatasetError>>
pub async fn send(
self
) -> Result<CreateDatasetOutput, SdkError<CreateDatasetError>>
Sends the request and returns the response.
If an error occurs, an SdkError will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
sourcepub fn dataset_name(self, input: impl Into<String>) -> Self
pub fn dataset_name(self, input: impl Into<String>) -> Self
A name for the dataset.
sourcepub fn set_dataset_name(self, input: Option<String>) -> Self
pub fn set_dataset_name(self, input: Option<String>) -> Self
A name for the dataset.
sourcepub fn domain(self, input: Domain) -> Self
pub fn domain(self, input: Domain) -> Self
The domain associated with the dataset. When you add a dataset to a dataset group, this value and the value specified for the Domain parameter of the CreateDatasetGroup operation must match.
The Domain and DatasetType that you choose determine the fields that must be present in the training data that you import to the dataset. For example, if you choose the RETAIL domain and TARGET_TIME_SERIES as the DatasetType, Amazon Forecast requires item_id, timestamp, and demand fields to be present in your data. For more information, see Importing datasets.
sourcepub fn set_domain(self, input: Option<Domain>) -> Self
pub fn set_domain(self, input: Option<Domain>) -> Self
The domain associated with the dataset. When you add a dataset to a dataset group, this value and the value specified for the Domain parameter of the CreateDatasetGroup operation must match.
The Domain and DatasetType that you choose determine the fields that must be present in the training data that you import to the dataset. For example, if you choose the RETAIL domain and TARGET_TIME_SERIES as the DatasetType, Amazon Forecast requires item_id, timestamp, and demand fields to be present in your data. For more information, see Importing datasets.
sourcepub fn dataset_type(self, input: DatasetType) -> Self
pub fn dataset_type(self, input: DatasetType) -> Self
The dataset type. Valid values depend on the chosen Domain.
sourcepub fn set_dataset_type(self, input: Option<DatasetType>) -> Self
pub fn set_dataset_type(self, input: Option<DatasetType>) -> Self
The dataset type. Valid values depend on the chosen Domain.
sourcepub fn data_frequency(self, input: impl Into<String>) -> Self
pub fn data_frequency(self, input: impl Into<String>) -> Self
The frequency of data collection. This parameter is required for RELATED_TIME_SERIES datasets.
Valid intervals are Y (Year), M (Month), W (Week), D (Day), H (Hour), 30min (30 minutes), 15min (15 minutes), 10min (10 minutes), 5min (5 minutes), and 1min (1 minute). For example, "D" indicates every day and "15min" indicates every 15 minutes.
sourcepub fn set_data_frequency(self, input: Option<String>) -> Self
pub fn set_data_frequency(self, input: Option<String>) -> Self
The frequency of data collection. This parameter is required for RELATED_TIME_SERIES datasets.
Valid intervals are Y (Year), M (Month), W (Week), D (Day), H (Hour), 30min (30 minutes), 15min (15 minutes), 10min (10 minutes), 5min (5 minutes), and 1min (1 minute). For example, "D" indicates every day and "15min" indicates every 15 minutes.
sourcepub fn schema(self, input: Schema) -> Self
pub fn schema(self, input: Schema) -> Self
The schema for the dataset. The schema attributes and their order must match the fields in your data. The dataset Domain and DatasetType that you choose determine the minimum required fields in your training data. For information about the required fields for a specific dataset domain and type, see Dataset Domains and Dataset Types.
sourcepub fn set_schema(self, input: Option<Schema>) -> Self
pub fn set_schema(self, input: Option<Schema>) -> Self
The schema for the dataset. The schema attributes and their order must match the fields in your data. The dataset Domain and DatasetType that you choose determine the minimum required fields in your training data. For information about the required fields for a specific dataset domain and type, see Dataset Domains and Dataset Types.
sourcepub fn encryption_config(self, input: EncryptionConfig) -> Self
pub fn encryption_config(self, input: EncryptionConfig) -> Self
An AWS Key Management Service (KMS) key and the AWS Identity and Access Management (IAM) role that Amazon Forecast can assume to access the key.
sourcepub fn set_encryption_config(self, input: Option<EncryptionConfig>) -> Self
pub fn set_encryption_config(self, input: Option<EncryptionConfig>) -> Self
An AWS Key Management Service (KMS) key and the AWS Identity and Access Management (IAM) role that Amazon Forecast can assume to access the key.
Appends an item to Tags.
To override the contents of this collection use set_tags.
The optional metadata that you apply to the dataset to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define.
The following basic restrictions apply to tags:
-
Maximum number of tags per resource - 50.
-
For each resource, each tag key must be unique, and each tag key can have only one value.
-
Maximum key length - 128 Unicode characters in UTF-8.
-
Maximum value length - 256 Unicode characters in UTF-8.
-
If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
-
Tag keys and values are case sensitive.
-
Do not use
aws:,AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value hasawsas its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix ofawsdo not count against your tags per resource limit.
The optional metadata that you apply to the dataset to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define.
The following basic restrictions apply to tags:
-
Maximum number of tags per resource - 50.
-
For each resource, each tag key must be unique, and each tag key can have only one value.
-
Maximum key length - 128 Unicode characters in UTF-8.
-
Maximum value length - 256 Unicode characters in UTF-8.
-
If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
-
Tag keys and values are case sensitive.
-
Do not use
aws:,AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value hasawsas its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix ofawsdo not count against your tags per resource limit.
Trait Implementations
sourceimpl Clone for CreateDataset
impl Clone for CreateDataset
sourcefn clone(&self) -> CreateDataset
fn clone(&self) -> CreateDataset
1.0.0 · sourcefn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source. Read more