paws.analytics (version 0.1.0)

datapipeline: AWS Data Pipeline

Description

AWS Data Pipeline configures and manages a data-driven workflow called a pipeline. AWS Data Pipeline handles the details of scheduling and ensuring that data dependencies are met so that your application can focus on processing the data.

AWS Data Pipeline provides a JAR implementation of a task runner called AWS Data Pipeline Task Runner. AWS Data Pipeline Task Runner provides logic for common data management scenarios, such as performing database queries and running data analysis using Amazon Elastic MapReduce (Amazon EMR). You can use AWS Data Pipeline Task Runner as your task runner, or you can write your own task runner to provide custom data management.

AWS Data Pipeline implements two main sets of functionality. Use the first set to create a pipeline and define data sources, schedules, dependencies, and the transforms to be performed on the data. Use the second set in your task runner application to receive the next task ready for processing. The logic for performing the task, such as querying the data, running data analysis, or converting the data from one format to another, is contained within the task runner. The task runner performs the task assigned to it by the web service, reporting progress to the web service as it does so. When the task is done, the task runner reports the final success or failure of the task to the web service.

Usage

datapipeline()

Arguments

Operations

activate_pipeline Validates the specified pipeline and starts processing pipeline tasks
add_tags Adds or modifies tags for the specified pipeline
create_pipeline Creates a new, empty pipeline
deactivate_pipeline Deactivates the specified running pipeline
delete_pipeline Deletes a pipeline, its pipeline definition, and its run history
describe_objects Gets the object definitions for a set of objects associated with the pipeline
describe_pipelines Retrieves metadata about one or more pipelines
evaluate_expression Task runners call EvaluateExpression to evaluate a string in the context of the specified object
get_pipeline_definition Gets the definition of the specified pipeline
list_pipelines Lists the pipeline identifiers for all active pipelines that you have permission to access
poll_for_task Task runners call PollForTask to receive a task to perform from AWS Data Pipeline
put_pipeline_definition Adds tasks, schedules, and preconditions to the specified pipeline
query_objects Queries the specified pipeline for the names of objects that match the specified set of conditions
remove_tags Removes existing tags from the specified pipeline
report_task_progress Task runners call ReportTaskProgress when assigned a task to acknowledge that it has the task
report_task_runner_heartbeat Task runners call ReportTaskRunnerHeartbeat every 15 minutes to indicate that they are operational
set_status Requests that the status of the specified physical or logical pipeline objects be updated in the specified pipeline
set_task_status Task runners call SetTaskStatus to notify AWS Data Pipeline that a task is completed and provide information about the final status

Examples

Run this code
# NOT RUN {
svc <- datapipeline()
svc$activate_pipeline(
  Foo = 123
)
# }
# NOT RUN {
# }

Run the code above in your browser using DataCamp Workspace