Environment
Pipelines have environments that are used to pass settings to your data source, data stores and other plugins.
Viewing your Environment
You can see what environment a data import is running with by going to that import, clicking the expand arrow and choosing the Environment tab. This tab will show all setting being used for your data source, data store and any other plugins you are using.
Some of these setting will be hidden with the value ***
, but when you copy and paste your environment, which you can do just by clicking anywhere in the environment window, these values will be shown.
Default Environment
Your data import’s environment by default will contain:
- Your data source’s settings
- Your data store’s settings
- dbt settings (if your data import uses dbt)
- A
MELTANO_DATABASE_URI
The MELTANO_DATABASE_URI
is where your workspaces’s Meltano project job information is stored. Also stored in along with the Meltano job information is any saved state. Depending on if a data source supports it, the saved state can be used in the next run of the data source as a starting point of where to get data from. This can save a time by limiting the amount of data synced to only new data since the state checkpoint.
By default we set every data import’s MELTANO_DATABASE_URI
to be the public
schema of your workspace’s default managed data store. (A postgres database that we set up and provide with each workspace). For more information about MELTANO_DATABASE_URI
s and State, see the links at the bottom of this page.
Editing Your Environment
You can add to or overwrite your environment variables by using a custom data import script. How to use and whats expected when you do use a custom data import script can also be found here.
In your custom data import script you can add new or overwrite existing environment variables with a single line:
EXPORT <NEW_OR_EXISTING_SETTING_NAME>=<NEW_VALUE>