#[non_exhaustive]
pub struct ContinuousExportDescription { pub export_id: Option<String>, pub status: Option<ContinuousExportStatus>, pub status_detail: Option<String>, pub s3_bucket: Option<String>, pub start_time: Option<DateTime>, pub stop_time: Option<DateTime>, pub data_source: Option<DataSource>, pub schema_storage_config: Option<HashMap<String, String>>, }
Expand description

A list of continuous export descriptions.

Fields (Non-exhaustive)§

This struct is marked as non-exhaustive
Non-exhaustive structs could have additional fields added in future. Therefore, non-exhaustive structs cannot be constructed in external crates using the traditional Struct { .. } syntax; cannot be matched against without a wildcard ..; and struct update syntax will not work.
§export_id: Option<String>

The unique ID assigned to this export.

§status: Option<ContinuousExportStatus>

Describes the status of the export. Can be one of the following values:

  • START_IN_PROGRESS - setting up resources to start continuous export.

  • START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.

  • ACTIVE - data is being exported to the customer bucket.

  • ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.

  • STOP_IN_PROGRESS - stopping the export.

  • STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.

  • INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.

§status_detail: Option<String>

Contains information about any errors that have occurred. This data type can have the following values:

  • ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.

  • DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.

  • FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.

  • FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.

  • INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.

  • LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.

    You can use one of the following two ways to resolve this issue.

    1. If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.

    2. You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.

      1. AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.

      2. AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.

  • S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.

  • S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.

§s3_bucket: Option<String>

The name of the s3 bucket where the export data parquet files are stored.

§start_time: Option<DateTime>

The timestamp representing when the continuous export was started.

§stop_time: Option<DateTime>

The timestamp that represents when this continuous export was stopped.

§data_source: Option<DataSource>

The type of data collector used to gather this data (currently only offered for AGENT).

§schema_storage_config: Option<HashMap<String, String>>

An object which describes how the data is stored.

  • databaseName - the name of the Glue database used to store the schema.

Implementations§

source§

impl ContinuousExportDescription

source

pub fn export_id(&self) -> Option<&str>

The unique ID assigned to this export.

source

pub fn status(&self) -> Option<&ContinuousExportStatus>

Describes the status of the export. Can be one of the following values:

  • START_IN_PROGRESS - setting up resources to start continuous export.

  • START_FAILED - an error occurred setting up continuous export. To recover, call start-continuous-export again.

  • ACTIVE - data is being exported to the customer bucket.

  • ERROR - an error occurred during export. To fix the issue, call stop-continuous-export and start-continuous-export.

  • STOP_IN_PROGRESS - stopping the export.

  • STOP_FAILED - an error occurred stopping the export. To recover, call stop-continuous-export again.

  • INACTIVE - the continuous export has been stopped. Data is no longer being exported to the customer bucket.

source

pub fn status_detail(&self) -> Option<&str>

Contains information about any errors that have occurred. This data type can have the following values:

  • ACCESS_DENIED - You don’t have permission to start Data Exploration in Amazon Athena. Contact your Amazon Web Services administrator for help. For more information, see Setting Up Amazon Web Services Application Discovery Service in the Application Discovery Service User Guide.

  • DELIVERY_STREAM_LIMIT_FAILURE - You reached the limit for Amazon Kinesis Data Firehose delivery streams. Reduce the number of streams or request a limit increase and try again. For more information, see Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide.

  • FIREHOSE_ROLE_MISSING - The Data Exploration feature is in an error state because your user is missing the Amazon Web ServicesApplicationDiscoveryServiceFirehose role. Turn on Data Exploration in Amazon Athena and try again. For more information, see Creating the Amazon Web ServicesApplicationDiscoveryServiceFirehose Role in the Application Discovery Service User Guide.

  • FIREHOSE_STREAM_DOES_NOT_EXIST - The Data Exploration feature is in an error state because your user is missing one or more of the Kinesis data delivery streams.

  • INTERNAL_FAILURE - The Data Exploration feature is in an error state because of an internal failure. Try again later. If this problem persists, contact Amazon Web Services Support.

  • LAKE_FORMATION_ACCESS_DENIED - You don't have sufficient lake formation permissions to start continuous export. For more information, see Upgrading Amazon Web Services Glue Data Permissions to the Amazon Web Services Lake Formation Model in the Amazon Web Services Lake Formation Developer Guide.

    You can use one of the following two ways to resolve this issue.

    1. If you don’t want to use the Lake Formation permission model, you can change the default Data Catalog settings to use only Amazon Web Services Identity and Access Management (IAM) access control for new databases. For more information, see Change Data Catalog Settings in the Lake Formation Developer Guide.

    2. You can give the service-linked IAM roles AWSServiceRoleForApplicationDiscoveryServiceContinuousExport and AWSApplicationDiscoveryServiceFirehose the required Lake Formation permissions. For more information, see Granting Database Permissions in the Lake Formation Developer Guide.

      1. AWSServiceRoleForApplicationDiscoveryServiceContinuousExport - Grant database creator permissions, which gives the role database creation ability and implicit permissions for any created tables. For more information, see Implicit Lake Formation Permissions in the Lake Formation Developer Guide.

      2. AWSApplicationDiscoveryServiceFirehose - Grant describe permissions for all tables in the database.

  • S3_BUCKET_LIMIT_FAILURE - You reached the limit for Amazon S3 buckets. Reduce the number of S3 buckets or request a limit increase and try again. For more information, see Bucket Restrictions and Limitations in the Amazon Simple Storage Service Developer Guide.

  • S3_NOT_SIGNED_UP - Your account is not signed up for the Amazon S3 service. You must sign up before you can use Amazon S3. You can sign up at the following URL: https://aws.amazon.com/s3.

source

pub fn s3_bucket(&self) -> Option<&str>

The name of the s3 bucket where the export data parquet files are stored.

source

pub fn start_time(&self) -> Option<&DateTime>

The timestamp representing when the continuous export was started.

source

pub fn stop_time(&self) -> Option<&DateTime>

The timestamp that represents when this continuous export was stopped.

source

pub fn data_source(&self) -> Option<&DataSource>

The type of data collector used to gather this data (currently only offered for AGENT).

source

pub fn schema_storage_config(&self) -> Option<&HashMap<String, String>>

An object which describes how the data is stored.

  • databaseName - the name of the Glue database used to store the schema.

source§

impl ContinuousExportDescription

source

pub fn builder() -> ContinuousExportDescriptionBuilder

Creates a new builder-style object to manufacture ContinuousExportDescription.

Trait Implementations§

source§

impl Clone for ContinuousExportDescription

source§

fn clone(&self) -> ContinuousExportDescription

Returns a copy of the value. Read more
1.0.0 · source§

fn clone_from(&mut self, source: &Self)

Performs copy-assignment from source. Read more
source§

impl Debug for ContinuousExportDescription

source§

fn fmt(&self, f: &mut Formatter<'_>) -> Result

Formats the value using the given formatter. Read more
source§

impl PartialEq for ContinuousExportDescription

source§

fn eq(&self, other: &ContinuousExportDescription) -> bool

This method tests for self and other values to be equal, and is used by ==.
1.0.0 · source§

fn ne(&self, other: &Rhs) -> bool

This method tests for !=. The default implementation is almost always sufficient, and should not be overridden without very good reason.
source§

impl StructuralPartialEq for ContinuousExportDescription

Auto Trait Implementations§

Blanket Implementations§

source§

impl<T> Any for T
where T: 'static + ?Sized,

source§

fn type_id(&self) -> TypeId

Gets the TypeId of self. Read more
source§

impl<T> Borrow<T> for T
where T: ?Sized,

source§

fn borrow(&self) -> &T

Immutably borrows from an owned value. Read more
source§

impl<T> BorrowMut<T> for T
where T: ?Sized,

source§

fn borrow_mut(&mut self) -> &mut T

Mutably borrows from an owned value. Read more
source§

impl<T> From<T> for T

source§

fn from(t: T) -> T

Returns the argument unchanged.

source§

impl<T> Instrument for T

source§

fn instrument(self, span: Span) -> Instrumented<Self>

Instruments this type with the provided Span, returning an Instrumented wrapper. Read more
source§

fn in_current_span(self) -> Instrumented<Self>

Instruments this type with the current Span, returning an Instrumented wrapper. Read more
source§

impl<T, U> Into<U> for T
where U: From<T>,

source§

fn into(self) -> U

Calls U::from(self).

That is, this conversion is whatever the implementation of From<T> for U chooses to do.

source§

impl<T> IntoEither for T

source§

fn into_either(self, into_left: bool) -> Either<Self, Self>

Converts self into a Left variant of Either<Self, Self> if into_left is true. Converts self into a Right variant of Either<Self, Self> otherwise. Read more
source§

fn into_either_with<F>(self, into_left: F) -> Either<Self, Self>
where F: FnOnce(&Self) -> bool,

Converts self into a Left variant of Either<Self, Self> if into_left(&self) returns true. Converts self into a Right variant of Either<Self, Self> otherwise. Read more
source§

impl<Unshared, Shared> IntoShared<Shared> for Unshared
where Shared: FromUnshared<Unshared>,

source§

fn into_shared(self) -> Shared

Creates a shared type from an unshared type.
source§

impl<T> Same for T

§

type Output = T

Should always be Self
source§

impl<T> ToOwned for T
where T: Clone,

§

type Owned = T

The resulting type after obtaining ownership.
source§

fn to_owned(&self) -> T

Creates owned data from borrowed data, usually by cloning. Read more
source§

fn clone_into(&self, target: &mut T)

Uses borrowed data to replace owned data, usually by cloning. Read more
source§

impl<T, U> TryFrom<U> for T
where U: Into<T>,

§

type Error = Infallible

The type returned in the event of a conversion error.
source§

fn try_from(value: U) -> Result<T, <T as TryFrom<U>>::Error>

Performs the conversion.
source§

impl<T, U> TryInto<U> for T
where U: TryFrom<T>,

§

type Error = <U as TryFrom<T>>::Error

The type returned in the event of a conversion error.
source§

fn try_into(self) -> Result<U, <U as TryFrom<T>>::Error>

Performs the conversion.
source§

impl<T> WithSubscriber for T

source§

fn with_subscriber<S>(self, subscriber: S) -> WithDispatch<Self>
where S: Into<Dispatch>,

Attaches the provided Subscriber to this type, returning a WithDispatch wrapper. Read more
source§

fn with_current_subscriber(self) -> WithDispatch<Self>

Attaches the current default Subscriber to this type, returning a WithDispatch wrapper. Read more