PipelineML
Reference documentation for orchestration actions that import and process data.
Use the Matatika pipeline YAML to orchestrate data actions in your workspace as code.
Pipeline definitions are stored as YAML file format, you can read more about the YAML format and its syntax here.
Example: pipelines/report_pipeline.yml
version: pipelines/v0.1
data_components:
- notebook
- sendgrid
actions:
- notebook:run-convert
- sendgrid:send
properties:
notebook.path: notebook/data_quality_report.ipynb
timeout: 1500
max_retries: 3
schedule: 0 0 0 * * 0
triggered_by:
- other-pipeline
- deploy
Key Information
Key | Datatype | Information |
---|---|---|
version |
String |
The version identifies this artifact type. |
data_components |
Array |
The meltano.yml data component name. |
actions |
Array |
The Meltano tasks that will be run as defined in your meltano.yml or Plugins. |
inline_script |
String |
Custom Bash script. Overrides actions if supplied. |
timeout |
Integer |
A timeout value in seconds that prevents pipelines from running for too long. A pipeline running longer that the timeout setting is automatically stopped by Matatika. |
max_retries |
Integer |
The maximum number of retries to attempt for a job ending with ERROR |
properties |
Map |
A map of properties, with Data Component name and setting as the key and the value e.g. data-component-name.setting=value , that configures the pipeline environment. |
schedule |
String |
The automated schedule for this pipeline, in a standard cron format with seconds. 0 0 9-17 * * MON-FRI on the hour nine-to-five weekdays. |
triggered_by |
Array of String |
Pipelines or workspace tasks that will trigger the pipeline on successful completion. Supported values for workspace tasks (case-insensitive):
|
Further Reading: