Expand description

AWS Data Pipeline configures and manages a data-driven workflow called a pipeline. AWS Data Pipeline handles the details of scheduling and ensuring that data dependencies are met so that your application can focus on processing the data.

AWS Data Pipeline provides a JAR implementation of a task runner called AWS Data Pipeline Task Runner. AWS Data Pipeline Task Runner provides logic for common data management scenarios, such as performing database queries and running data analysis using Amazon Elastic MapReduce (Amazon EMR). You can use AWS Data Pipeline Task Runner as your task runner, or you can write your own task runner to provide custom data management.

AWS Data Pipeline implements two main sets of functionality. Use the first set to create a pipeline and define data sources, schedules, dependencies, and the transforms to be performed on the data. Use the second set in your task runner application to receive the next task ready for processing. The logic for performing the task, such as querying the data, running data analysis, or converting the data from one format to another, is contained within the task runner. The task runner performs the task assigned to it by the web service, reporting progress to the web service as it does so. When the task is done, the task runner reports the final success or failure of the task to the web service.

Crate Organization

The entry point for most customers will be Client. Client exposes one method for each API offered by the service.

Some APIs require complex or nested arguments. These exist in model.

Lastly, errors that can be returned by the service are contained within error. Error defines a meta error encompassing all possible errors that can be returned by the service.

The other modules within this crate are not required for normal usage.

Modules

Client and fluent builders for calling the service.

Configuration for the service.

Errors that can occur when calling the service.

Input structures for operations.

Generated accessors for nested fields

Base Middleware Stack

Data structures used by operation inputs/outputs.

All operations that this crate can perform.

Output structures for operations.

Paginators for the service

Re-exported types from supporting crates.

Structs

App name that can be configured with an AWS SDK client to become part of the user agent string.

Client for AWS Data Pipeline

Service config.

AWS SDK Credentials

API Endpoint

The region to send requests to.

Retry configuration for requests.

Enums

All possible error types for this service.

Statics

Crate version number.