Pipelines
A pipeline defines a set of runnable actions composed from datacomponents to complete a set of tasks - for example, ELT. Pipelines are run as jobs, either manually or on a predetermined schedule. Only a single pipeline can be run at any given time.
Objects
Pipeline
Path | Type | Format | Description |
---|---|---|---|
id |
String |
Version 4 UUID | The pipeline ID |
status |
String |
Pipeline Status | |
name |
String |
The pipeline name | |
schedule |
String |
Cron | The interval at which to launch a new job e.g. 0 0 9-17 * * MON-FRI launches a job on the hour nine-to-five weekdays |
timeout |
Integer |
Unsigned | The number of seconds after which the job will terminate - if set to 0 , an implicit default value of 300 seconds is used |
maxRetries |
Integer |
Unsigned | The maximum number of retries to attempt for a job ending with ERROR |
script |
String |
Bash script | Custom script to execute during a job |
created |
String |
ISO 8601 timestamp | When the pipeline was created |
lastModified |
String |
ISO 8601 timestamp | When the pipeline was last modified |
properties |
Properties |
The pipeline properties, defined by the dataplugin settings of each datacomponent |
|
dataComponents |
Array of String |
Array of Datacomponent name s |
The pipeline datacomponent name s or create / update with dataplugin fullyQualifiedName |
actions |
Array of String |
Array of Datacomponent name s or commands |
The pipeline actions to run during a job |
triggeredBy |
Array of String |
Array of pipeline name s or workspace task identifiers |
Pipelines or workspace tasks that will trigger the pipeline on successful completion Supported values for workspace tasks (case-insensitive):
|
{
"id" : "e211e8fe-80b9-43b5-8941-de005fbdb919",
"status" : "READY",
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.840956",
"lastModified" : "2024-12-16T10:02:09.840957",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
},
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/verification",
"type" : "POST"
},
"create job" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "POST"
}
}
}
Properties
For each setting s
in the datacomponents’ dataplugin settings
for each
Path | Type | Description |
---|---|---|
s.name |
s.kind |
Refer to s.description |
- Any required settings not satisfied by a datacomponent property must be provided as a pipeline property
- Any settings that are already satisfied by a datacomponent property can be overridden by a pipeline property
Formats
Pipeline Status
Value | Description |
---|---|
READY |
The pipeline completed processing resource changes |
PROVISIONING |
The pipeline is processing resource changes |
FAILED |
The pipeline failed to process resource changes |
Requests
- View all pipelines in a workspace
- View a pipeline
- Initialise a pipeline in a workspace
- Create or update a pipeline in a workspace
- Create or update a pipeline as a draft
- Validate a pipeline configuration in a workspace
- Verify a pipeline
- Delete a pipeline
- View pipeline metrics
See Also
- View all running or completed jobs for a pipeline
- Create a job from a pipeline
- Subscribe to a pipeline
View all pipelines in a workspace
GET
/api/workspaces/{workspace-id}/pipelines
Returns all configured pipelines in the workspace {workspace-id}
.
Prerequisites
- Workspace
{workspace-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines' -i -X GET \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("GET", url, headers=headers)
print(response.text.encode('utf8'))
Response
200 OK
Pipeline collection with HAL links.
{
"_embedded" : {
"pipelines" : [ {
"id" : "b3c66256-386a-4156-a59d-d92d464065d4",
"status" : "DRAFT",
"name" : "SIT-generated pipeline [2024-12-16T10:02:08.781891]",
"schedule" : "0 0 0 25 12 ?",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.141541",
"lastModified" : "2024-12-16T10:02:09.141542",
"properties" : { },
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/verification",
"type" : "POST"
}
}
}, {
"id" : "e211e8fe-80b9-43b5-8941-de005fbdb919",
"status" : "READY",
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.840956",
"lastModified" : "2024-12-16T10:02:09.840957",
"properties" : { },
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/verification",
"type" : "POST"
},
"create job" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "POST"
}
}
} ]
},
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines?page=0&size=20&sort=name,asc"
}
},
"page" : {
"size" : 20,
"totalElements" : 2,
"totalPages" : 1,
"number" : 0
}
}
View a pipeline
GET
/api/pipelines/{pipeline-id}
Returns the pipeline {pipeline-id}
.
Prerequisites
- Pipeline
{pipeline-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919' -i -X GET \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("GET", url, headers=headers)
print(response.text.encode('utf8'))
Response
200 OK
Pipeline with HAL links.
{
"id" : "e211e8fe-80b9-43b5-8941-de005fbdb919",
"status" : "READY",
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.840956",
"lastModified" : "2024-12-16T10:02:09.840957",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
},
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/verification",
"type" : "POST"
},
"create job" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "POST"
}
}
}
Initialise a pipeline in a workspace
POST
/api/workspaces/{workspace-id}/pipelines
Initialises a new pipeline in the workspace {workspace-id}
.
Prerequisites
- Workspace
{workspace-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines' -i -X POST \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("POST", url, headers=headers)
print(response.text.encode('utf8'))
Response
200 OK
Pipeline with HAL links.
{
"id" : "b3c66256-386a-4156-a59d-d92d464065d4",
"status" : "PROVISIONING",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:08.720309653",
"lastModified" : "2024-12-16T10:02:08.720310053",
"properties" : { },
"dataComponents" : [ ],
"actions" : [ ],
"triggeredBy" : [ ],
"_links" : {
"create pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4",
"type" : "PUT"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/draft",
"type" : "PUT"
},
"validate pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/validation",
"type" : "POST"
}
}
}
Create or update a pipeline in a workspace
PUT
/api/workspaces/{workspace-id}/pipelines/{pipeline-id}
Creates or updates the pipeline {pipeline-id}
in the workspace {workspace-id}
.
Prerequisites
- Workspace
{workspace-id}
must exist
Request
Body
Pipeline resource.
{
"name" : "SIT-generated pipeline [2024-12-16T10:02:08.781891]",
"dataComponents" : [ "extractors/tap-google-analytics", "Warehouse", "dbt" ],
"schedule" : "0 0 0 25 12 ?",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
}
}
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4' -i -X PUT \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json' \
-d '{
"name" : "SIT-generated pipeline [2024-12-16T10:02:08.781891]",
"dataComponents" : [ "extractors/tap-google-analytics", "Warehouse", "dbt" ],
"schedule" : "0 0 0 25 12 ?",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
}
}'
Python (requests
)
import requests
url = "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4"
data = {
"name" : "SIT-generated pipeline [2024-12-16T10:02:08.781891]",
"dataComponents" : [ "extractors/tap-google-analytics", "Warehouse", "dbt" ],
"schedule" : "0 0 0 25 12 ?",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
}
}
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("PUT", url, headers=headers, data=data)
print(response.text.encode('utf8'))
Response
200 OK / 201 Created
Pipeline with HAL links.
{
"id" : "b3c66256-386a-4156-a59d-d92d464065d4",
"status" : "READY",
"name" : "SIT-generated pipeline [2024-12-16T10:02:08.781891]",
"schedule" : "0 0 0 25 12 ?",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.141541",
"lastModified" : "2024-12-16T10:02:09.141542",
"properties" : {
"tap-google-analytics.view_id" : "1234567890",
"tap-google-analytics.oauth_credentials.client_secret" : "client_secret",
"tap-google-analytics.reports" : "reports",
"tap-google-analytics.oauth_credentials.refresh_token" : "refresh_token",
"tap-google-analytics.start_date" : "2024-11-16T10:02:08.788286Z",
"tap-google-analytics.end_date" : "2024-12-16T10:02:08.788341Z",
"tap-google-analytics.oauth_credentials.client_id" : "client_id",
"tap-google-analytics.oauth_credentials.access_token" : "access_token"
},
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/verification",
"type" : "POST"
},
"create job" : {
"href" : "https://app.matatika.com/api/pipelines/b3c66256-386a-4156-a59d-d92d464065d4/jobs",
"type" : "POST"
}
}
}
Create or update a pipeline as a draft
PUT
/api/workspaces/{workspace-id}/pipelines/{pipeline-id}/draft
Creates or updates the pipeline {pipeline-id}
in the workspace {workspace-id}
as a draft.
Prerequisites
- Workspace
{workspace-id}
must exist
Request
Body
Pipeline resource.
{
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162]",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft' -i -X PUT \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json' \
-d '{
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162]",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}'
Python (requests
)
import requests
url = "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft"
data = {
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162]",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("PUT", url, headers=headers, data=data)
print(response.text.encode('utf8'))
Response
200 OK / 201 Created
Pipeline with HAL links.
{
"id" : "e211e8fe-80b9-43b5-8941-de005fbdb919",
"status" : "DRAFT",
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162]",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T10:02:09.840956",
"lastModified" : "2024-12-16T10:02:09.840957",
"properties" : { },
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"_embedded" : {
"dataComponents" : [ {
"id" : "842a3419-b67d-46d5-a1be-8e8442e0e6e4",
"created" : "2024-12-16T10:02:08.953528",
"lastModified" : "2024-12-16T10:02:08.953528",
"name" : "tap-google-analytics",
"properties" : { },
"commands" : { },
"dataPlugin" : "extractors/tap-google-analytics--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "618281ff-6dc8-4b92-b4f1-bff318f893af",
"pluginType" : "EXTRACTOR",
"name" : "tap-google-analytics",
"namespace" : "tap_google_analytics",
"variant" : "matatika",
"label" : "Google Analytics",
"description" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.",
"logoUrl" : "/assets/images/datasource/tap-google-analytics.svg",
"hidden" : false,
"docs" : "https://www.matatika.com/docs/instant-insights/tap-google-analytics/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/tap-google-analytics",
"capabilities" : [ "DISCOVER", "CATALOG", "STATE" ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "oauth_credentials.authorization_url",
"aliases" : [ ],
"label" : "OAuth identity provider authorization endpoint used create and refresh tokens",
"value" : "https://oauth2.googleapis.com/token",
"options" : [ ],
"kind" : "STRING",
"description" : "The endpoint used to create and refresh OAuth tokens.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.scope",
"aliases" : [ ],
"label" : "OAuth scopes we need to request access to",
"value" : "profile email https://www.googleapis.com/auth/analytics.readonly",
"options" : [ ],
"kind" : "STRING",
"description" : "The specific scopes we need to request access to in order to connect to Google Analytics.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.access_token",
"aliases" : [ ],
"label" : "Access Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to authenticate and authorize API requests.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_token",
"aliases" : [ ],
"label" : "OAuth Refresh Token",
"options" : [ ],
"kind" : "STRING",
"description" : "The token used to refresh the access token when it expires.",
"hidden" : true,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url",
"aliases" : [ ],
"label" : "Optional - will be called with 'oauth_credentials.refresh_token' to refresh the access token",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional function that will be called to refresh the access token using the refresh token.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.refresh_proxy_url_auth",
"aliases" : [ ],
"label" : "Optional - Sets Authorization header on 'oauth_credentials.refresh_url' request",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional setting that sets the Authorization header on the refresh URL request.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_id",
"aliases" : [ ],
"label" : "Optional - OAuth Client ID used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client ID used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "oauth_credentials.client_secret",
"aliases" : [ ],
"label" : "Optional - OAuth Client Secret used if refresh_proxy_url not supplied",
"options" : [ ],
"kind" : "STRING",
"description" : "An optional OAuth Client Secret used if the refresh proxy URL is not supplied.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "view_id",
"aliases" : [ ],
"label" : "View ID",
"options" : [ ],
"placeholder" : "Ex. 198343027",
"kind" : "STRING",
"description" : "The ID of the Google Analytics view to retrieve data from.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "reports",
"aliases" : [ ],
"label" : "Reports",
"options" : [ ],
"placeholder" : "Ex. my_report_definition.json",
"kind" : "STRING",
"description" : "The specific reports to retrieve data from in the Google Analytics view.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "start_date",
"aliases" : [ ],
"label" : "Start date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The start date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "end_date",
"aliases" : [ ],
"label" : "End date",
"options" : [ ],
"kind" : "DATE_ISO8601",
"description" : "The end date for the date range of data to retrieve.",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Google Analytics is a web analytics service that provides insights into website traffic and user behavior.\n\nGoogle Analytics allows website owners to track and analyze various metrics related to their website's performance, such as the number of visitors, pageviews, bounce rate, and average session duration. It also provides information on the demographics and interests of website visitors, as well as the sources of traffic, including organic search, paid search, social media, and referrals. This data can be used to optimize website content and marketing strategies, as well as to measure the effectiveness of advertising campaigns. Additionally, Google Analytics offers advanced features such as goal tracking, e-commerce tracking, and custom reporting, making it a powerful tool for businesses of all sizes.\n### Prerequisites\nTo obtain the OAuth identity provider authorization endpoint used to create and refresh tokens, you need to create a project in the Google API Console and enable the Google Analytics API. Then, you can create OAuth 2.0 credentials and configure the authorized redirect URIs. The authorization endpoint will be provided in the credentials.\n\nThe OAuth scopes you need to request access to depend on the specific data you want to access in Google Analytics. For example, if you want to read data from a specific view, you will need to request the \"https://www.googleapis.com/auth/analytics.readonly\" scope. You can find a list of available scopes in the Google Analytics API documentation.\n\nTo obtain the Access Token and OAuth Refresh Token, you need to authenticate the user and obtain their consent to access their Google Analytics data. This can be done using the Google Sign-In API or the OAuth 2.0 authorization flow. Once the user has granted access, you will receive an Access Token and a Refresh Token that you can use to make API requests.\n\nTo obtain the View ID, you need to log in to your Google Analytics account and navigate to the Admin section. From there, you can select the account, property, and view that you want to access and find the View ID in the View Settings.\n\n## Settings\n\n\n### View ID\n\nThe ID of the Google Analytics view to retrieve data from.\n\n### Reports\n\nThe specific reports to retrieve data from in the Google Analytics view.\n\n### Start date\n\nThe start date for the date range of data to retrieve.\n\n### End date\n\nThe end date for the date range of data to retrieve.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/618281ff-6dc8-4b92-b4f1-bff318f893af",
"type" : "PUT"
}
}
}
},
"managed" : false,
"draft" : true,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/842a3419-b67d-46d5-a1be-8e8442e0e6e4"
}
}
}, {
"id" : "020daf95-0458-43ac-907e-f5d7ba5bdbf6",
"created" : "2024-12-16T10:02:02.895019",
"lastModified" : "2024-12-16T10:02:02.89502",
"name" : "Warehouse",
"properties" : {
"password" : "IxBx93ih_huFT374K_wRVU19el",
"default_target_schema" : "analytics",
"dbname" : "dxdoili",
"port" : "5432",
"host" : "sharp-banana.postgres.database.azure.com",
"user" : "dxdoili"
},
"commands" : { },
"dataPlugin" : "loaders/target-postgres--matatika",
"_embedded" : {
"dataplugin" : {
"id" : "75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"pluginType" : "LOADER",
"name" : "target-postgres",
"namespace" : "postgres_transferwise",
"variant" : "matatika",
"label" : "Postgres Warehouse",
"description" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.",
"logoUrl" : "/assets/logos/loaders/postgres.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/target-postgres/",
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "git+https://github.com/Matatika/[email protected]",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "user",
"aliases" : [ "username" ],
"label" : "User",
"options" : [ ],
"kind" : "STRING",
"description" : "The username used to connect to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "password",
"aliases" : [ ],
"label" : "Password",
"options" : [ ],
"kind" : "STRING",
"description" : "The password used to authenticate the user.",
"hidden" : false,
"sensitive" : true,
"required" : "true",
"protected" : false
}, {
"name" : "host",
"aliases" : [ "address" ],
"label" : "Host",
"options" : [ ],
"kind" : "STRING",
"description" : "The hostname or IP address of the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "port",
"aliases" : [ ],
"label" : "Port",
"value" : "5432",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The port number used to connect to the Postgres Warehouse server.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "dbname",
"aliases" : [ "database" ],
"label" : "Database Name",
"options" : [ ],
"kind" : "STRING",
"description" : "The name of the database to connect to.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "default_target_schema",
"aliases" : [ ],
"label" : "Default Target Schema",
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"description" : "The default schema to use when writing data to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"required" : "true",
"protected" : false
}, {
"name" : "ssl",
"aliases" : [ ],
"label" : "SSL",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to use SSL encryption when connecting to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false,
"value_post_processor" : "STRINGIFY"
}, {
"name" : "batch_size_rows",
"aliases" : [ ],
"label" : "Batch Size Rows",
"value" : "100000",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of rows to write to the Postgres Warehouse in each batch.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "underscore_camel_case_fields",
"aliases" : [ ],
"label" : "Underscore Camel Case Fields",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to convert field names from camel case to underscore-separated format.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "flush_all_streams",
"aliases" : [ ],
"label" : "Flush All Streams",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to flush all streams to the Postgres Warehouse before closing the connection.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism",
"aliases" : [ ],
"label" : "Parallelism",
"value" : "0",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "parallelism_max",
"aliases" : [ ],
"label" : "Max Parallelism",
"value" : "16",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum number of threads to use when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "default_target_schema_select_permission",
"aliases" : [ ],
"label" : "Default Target Schema Select Permission",
"options" : [ ],
"kind" : "STRING",
"description" : "The permission level required to select data from the default target schema.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "schema_mapping",
"aliases" : [ ],
"label" : "Schema Mapping",
"options" : [ ],
"kind" : "STRING",
"description" : "A mapping of source schema names to target schema names.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "add_metadata_columns",
"aliases" : [ ],
"label" : "Add Metadata Columns",
"value" : "true",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to add metadata columns to the target table.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "hard_delete",
"aliases" : [ ],
"label" : "Hard Delete",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to perform hard deletes when deleting data from the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "data_flattening_max_level",
"aliases" : [ ],
"label" : "Data Flattening Max Level",
"value" : "10",
"options" : [ ],
"kind" : "INTEGER",
"description" : "The maximum level of nested data structures to flatten when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
}, {
"name" : "primary_key_required",
"aliases" : [ ],
"label" : "Primary Key Required",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not a primary key is required for the target table.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "validate_records",
"aliases" : [ ],
"label" : "Validate Records",
"value" : "false",
"options" : [ ],
"kind" : "BOOLEAN",
"description" : "Whether or not to validate records before writing them to the Postgres Warehouse.",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "temp_dir",
"aliases" : [ ],
"label" : "Temporary Directory",
"options" : [ ],
"kind" : "STRING",
"description" : "The directory to use for temporary files when writing data to the Postgres Warehouse.",
"hidden" : true,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : "Postgres Warehouse is a data warehousing solution built on top of the Postgres database management system.\n\nPostgres Warehouse is designed to handle large volumes of data and complex queries, making it an ideal solution for businesses that need to store and analyze large amounts of data. It provides a number of features that are specifically tailored to data warehousing, such as columnar storage, parallel processing, and support for advanced analytics. Additionally, Postgres Warehouse is highly scalable, allowing businesses to easily add more resources as their data needs grow. Overall, Postgres Warehouse is a powerful and flexible data warehousing solution that can help businesses make better decisions by providing them with the insights they need to succeed.\n### Prerequisites\nThe process of obtaining the required settings for connecting to a Postgres Warehouse may vary depending on the specific setup and configuration of the database. However, here are some general ways to obtain each of the required settings:\n\n- User: The user is typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the username.\n- Password: The password is also typically created when the database is set up. You can ask the database administrator or check the database documentation to find out the password.\n- Host: The host is the server where the database is located. You can ask the database administrator or check the database documentation to find out the host name or IP address.\n- Port: The port is the number that the database listens on for incoming connections. The default port for Postgres is 5432, but it may be different depending on the configuration. You can ask the database administrator or check the database documentation to find out the port number.\n- Database Name: The database name is the name of the specific database you want to connect to. You can ask the database administrator or check the database documentation to find out the database name.\n- Default Target Schema: The default target schema is the schema that you want to use as the default when connecting to the database. This may be set up by the database administrator or you may need to create it yourself. You can ask the database administrator or check the database documentation to find out the default target schema.\n\n## Settings\n\n\n### User\n\nThe username used to connect to the Postgres Warehouse.\n\n### Password\n\nThe password used to authenticate the user.\n\n### Host\n\nThe hostname or IP address of the Postgres Warehouse server.\n\n### Port\n\nThe port number used to connect to the Postgres Warehouse server.\n\n### Database Name\n\nThe name of the database to connect to.\n\n### Default Target Schema\n\nThe default schema to use when writing data to the Postgres Warehouse.\n\n### Batch Size Rows\n\nThe number of rows to write to the Postgres Warehouse in each batch.\n\n### Primary Key Required\n\nWhether or not a primary key is required for the target table.\n\n### Validate Records\n\nWhether or not to validate records before writing them to the Postgres Warehouse.",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/75083d9e-2496-4cfc-ba38-c04ef593fdd1",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/020daf95-0458-43ac-907e-f5d7ba5bdbf6"
}
}
}, {
"id" : "876f7737-3aea-4958-accd-3dc318d044f2",
"created" : "2024-12-16T10:02:02.961373",
"lastModified" : "2024-12-16T10:02:02.961374",
"name" : "dbt",
"properties" : { },
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"dataPlugin" : "transformers/dbt--dbt-labs",
"_embedded" : {
"dataplugin" : {
"id" : "fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"pluginType" : "TRANSFORMER",
"name" : "dbt",
"namespace" : "dbt",
"variant" : "dbt-labs",
"label" : "dbt",
"description" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"logoUrl" : "/assets/images/transformer/dbt.png",
"hidden" : false,
"docs" : "https://www.matatika.com/data-details/dbt/",
"pipUrl" : "dbt-core~=1.3.0 dbt-postgres~=1.3.0 dbt-snowflake~=1.3.0\n",
"repo" : "https://github.com/dbt-labs/dbt-core",
"capabilities" : [ ],
"select" : [ ],
"update" : { },
"vars" : { },
"settings" : [ {
"name" : "project_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "profiles_dir",
"aliases" : [ ],
"value" : "$MELTANO_PROJECT_ROOT/transform/profile",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"env" : "DBT_PROFILES_DIR",
"protected" : false
}, {
"name" : "target",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__DIALECT",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "source_schema",
"aliases" : [ ],
"value" : "$MELTANO_LOAD__TARGET_SCHEMA",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "target_schema",
"aliases" : [ ],
"value" : "analytics",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
}, {
"name" : "models",
"aliases" : [ ],
"value" : "$MELTANO_TRANSFORM__PACKAGE_NAME $MELTANO_EXTRACTOR_NAMESPACE my_meltano_project",
"options" : [ ],
"kind" : "STRING",
"hidden" : false,
"sensitive" : false,
"protected" : false
} ],
"variants" : [ ],
"commands" : {
"compile" : {
"args" : "compile",
"description" : "Generates executable SQL from source model, test, and analysis files. Compiled SQL files are written to the target/ directory."
},
"seed" : {
"args" : "seed",
"description" : "Load data from csv files into your data warehouse."
},
"test" : {
"args" : "test",
"description" : "Runs tests on data in deployed models."
},
"docs-generate" : {
"args" : "docs generate",
"description" : "Generate documentation artifacts for your project."
},
"deps" : {
"args" : "deps",
"description" : "Pull the most recent version of the dependencies listed in packages.yml"
},
"run" : {
"args" : "run",
"description" : "Compile SQL and execute against the current target database."
},
"clean" : {
"args" : "clean",
"description" : "Delete all folders in the clean-targets list (usually the dbt_modules and target directories.)"
},
"snapshot" : {
"args" : "snapshot",
"description" : "Execute snapshots defined in your project."
}
},
"matatikaHidden" : false,
"requires" : [ {
"id" : "e6c1ad3d-ebf5-4c4a-b129-f68156b47555",
"pluginType" : "FILE",
"name" : "files-dbt",
"namespace" : "dbt",
"variant" : "matatika",
"description" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n",
"hidden" : false,
"pipUrl" : "git+https://github.com/Matatika/[email protected]",
"repo" : "https://github.com/Matatika/files-dbt",
"capabilities" : [ ],
"select" : [ ],
"update" : {
"transform/profile/profiles.yml" : "true"
},
"vars" : { },
"settings" : [ ],
"variants" : [ ],
"commands" : { },
"matatikaHidden" : false,
"requires" : [ ],
"fullDescription" : " Files dbt is a file bundle that automatically configures your project to run transforms with dbt.\nThe bundle includes template project configuration:\n\n- transform/models (directory)\n- transform/profile/profiles.yml\n- transform/dbt_project.yml\n- transform/.gitignore\n- transform/macros/centralize_test_failures.sql\n"
} ],
"fullDescription" : " Power your project transformations with dbt™, a SQL-first transformation tool that enables analytics engineers to develop transformations with code.\n\n***Version Control and CI/CD***\n\nUse Matatika to deploy and promote changes between dev, UAT, and production environments.\n\n***Test and Document***\n\nUse Matatika to develop and test every model prior to production release, and share dynamically generated documentation with all stakeholders.\n\n***Develop***\n\nWrite modular data transformations in .sql – Matatika together with dbt handles the chore of dependency management. ",
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd"
},
"update dataplugin" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/dataplugins/fb3ebc88-708b-44e8-b1bf-bf6a03c8fecd",
"type" : "PUT"
}
}
}
},
"managed" : true,
"draft" : false,
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"update datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
},
"delete datacomponent" : {
"href" : "https://app.matatika.com/api/datacomponents/876f7737-3aea-4958-accd-3dc318d044f2"
}
}
} ]
},
"_links" : {
"update pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "PUT"
},
"delete pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919",
"type" : "DELETE"
},
"draft pipeline" : {
"href" : "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/draft",
"type" : "PUT"
},
"self" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
},
"environment" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/environment"
},
"jobs" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/jobs",
"type" : "GET"
},
"metrics" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/metrics"
},
"add subscription" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/subscriptions"
},
"verify pipeline" : {
"href" : "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919/verification",
"type" : "POST"
}
}
}
Validate a pipeline configuration in a workspace
POST
/api/workspaces/{workspace-id}/pipelines/validation
Validates a pipeline configuration in the workspace {workspace-id}
.
Prerequisites
- Workspace
{workspace-id}
must exist
Request
Body
Pipeline resource.
{
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/validation' -i -X POST \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json' \
-d '{
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}'
Python (requests
)
import requests
url = "https://app.matatika.com/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/validation"
data = {
"name" : "SIT-generated pipeline [2024-12-16T10:02:09.603162] (updated)",
"dataComponents" : [ "tap-google-analytics", "Warehouse", "dbt" ]
}
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("POST", url, headers=headers, data=data)
print(response.text.encode('utf8'))
Response
200 OK
No response body provided.
400 Bad Request
Pipeline property validation errors.
{
"timestamp" : "2024-12-16T10:02:12.775218966",
"status" : 400,
"error" : "Bad Request",
"message" : "3 validation errors from 'resource'",
"errors" : [ {
"codes" : [ "NotBlank.oauth_credentials.access_token", "NotBlank" ],
"defaultMessage" : "No value given for setting",
"objectName" : "resource",
"field" : "properties.tap-google-analytics.oauth_credentials.access_token",
"bindingFailure" : true,
"code" : "NotBlank"
}, {
"codes" : [ "NotBlank.oauth_credentials.refresh_token", "NotBlank" ],
"defaultMessage" : "No value given for setting",
"objectName" : "resource",
"field" : "properties.tap-google-analytics.oauth_credentials.refresh_token",
"bindingFailure" : true,
"code" : "NotBlank"
}, {
"codes" : [ "NotBlank.view_id", "NotBlank" ],
"defaultMessage" : "No value given for setting",
"objectName" : "resource",
"field" : "properties.tap-google-analytics.view_id",
"bindingFailure" : true,
"code" : "NotBlank"
} ],
"path" : "/api/workspaces/f1a081e0-bab9-4775-86fb-b77ce993e84d/pipelines/validation"
}
Verify a pipeline
POST
/api/pipelines/{pipeline-id}/verification
Verifies the configuration of the pipeline {pipeline-id}
.
Prerequisites
- Pipeline
{pipeline-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/pipelines/2180a2e8-f75f-4319-8a91-f47046535967/verification' -i -X POST \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/pipelines/2180a2e8-f75f-4319-8a91-f47046535967/verification"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("POST", url, headers=headers)
print(response.text.encode('utf8'))
Response
200 OK
Job with HAL links.
{
"id" : "886c8e69-121b-4c4d-a275-62a1b942f01d",
"created" : "2024-12-16T09:56:08.021023",
"type" : "PIPELINE_VERIFY",
"maxAttempts" : 0,
"attempt" : 0,
"status" : "QUEUED",
"_embedded" : {
"pipeline" : {
"id" : "2180a2e8-f75f-4319-8a91-f47046535967",
"status" : "READY",
"name" : "SIT-Generated Pipeline [2024-12-16T09:54:29.465456]",
"timeout" : 0,
"maxRetries" : 0,
"created" : "2024-12-16T09:54:29.753328",
"lastModified" : "2024-12-16T09:54:29.753328",
"properties" : { },
"dataComponents" : [ "tap-matatika-sit", "Warehouse", "dbt" ],
"actions" : [ ],
"triggeredBy" : [ ],
"repositoryPath" : "pipelines/SIT-Generated Pipeline [2024-12-16T09:54:29.465456].yml"
},
"profile" : {
"id" : "auth0|5eb0327cbfd7490bff55feeb",
"name" : "[email protected]",
"handle" : "@sit+prod",
"email" : "[email protected]"
}
},
"_links" : {
"self" : {
"href" : "https://app.matatika.com/api/jobs/886c8e69-121b-4c4d-a275-62a1b942f01d"
},
"delete job" : {
"href" : "https://app.matatika.com/api/jobs/886c8e69-121b-4c4d-a275-62a1b942f01d",
"type" : "DELETE"
},
"logs" : {
"href" : "https://app.matatika.com/api/jobs/886c8e69-121b-4c4d-a275-62a1b942f01d/logs?sequence=0",
"type" : "GET"
},
"withdraw job" : {
"href" : "https://app.matatika.com/api/jobs/886c8e69-121b-4c4d-a275-62a1b942f01d/stopped",
"type" : "PUT"
}
}
}
Delete a pipeline
DELETE
/api/pipelines/{pipeline-id}
Deletes the pipeline {pipeline-id}
.
Prerequisites
- Pipeline
{pipeline-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919' -i -X DELETE \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/pipelines/e211e8fe-80b9-43b5-8941-de005fbdb919"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("DELETE", url, headers=headers)
print(response.text.encode('utf8'))
Response
204 No Content
No response body provided.
View pipeline metrics
GET
/api/pipelines/{pipeline-id}/metrics
Returns the pipeline metrics for each job of {pipeline-id}
.
Prerequisites
- Pipeline
{pipeline-id}
must exist
Request
Example Snippets
cURL
curl -H "Authorization: Bearer $ACCESS_TOKEN" 'https://app.matatika.com/api/pipelines/2180a2e8-f75f-4319-8a91-f47046535967/metrics' -i -X GET \
-H 'Accept: application/json, application/javascript, text/javascript, text/json' \
-H 'Content-Type: application/json'
Python (requests
)
import requests
url = "https://app.matatika.com/api/pipelines/2180a2e8-f75f-4319-8a91-f47046535967/metrics"
headers = {
'Authorization': ACCESS_TOKEN
}
response = requests.request("GET", url, headers=headers)
print(response.text.encode('utf8'))
Response
200
: The dataset data (defaults to JSON format).
[ {
"metrics.job-created" : "2024-12-16 09:56:50",
"metrics.value" : 6.0
} ]
204
: No response body, metrics not enabled.