Amplitude Source dbt Package (Docs)
What does this dbt package do?
- Materializes Amplitude staging tables which leverage data in the format described by this ERD. These staging tables clean, test, and prepare your Amplitude data from Fivetran's connector for analysis by doing the following:
- Name columns for consistency across all packages and for easier analysis
- Adds freshness tests to source data
- Adds column-level testing where applicable. For example, all primary keys are tested for uniqueness and non-null values.
- Generates a comprehensive data dictionary of your Amplitude data through the dbt docs site.
- These tables are designed to work simultaneously with our Amplitude transformation package.
- Refer to our Docs site for more details about these materialized models.
How do I use the dbt package?
Step 1: Prerequisites
To use this dbt package, you must have the following:
- At least one Fivetran Amplitude connector syncing data into your destination.
- A BigQuery, Snowflake, Redshift, PostgreSQL, or Databricks destination.
Databricks dispatch configuration
If you are using a Databricks destination with this package, you must add the following (or a variation of the following) dispatch configuration within your dbt_project.yml
. This is required in order for the package to accurately search for macros within the dbt-labs/spark_utils
then the dbt-labs/dbt_utils
packages respectively.
dispatch:
- macro_namespace: dbt_utils
search_order: ['spark_utils', 'dbt_utils']
Step 2: Install the package (skip if also using the amplitude
transformation package)
If you are not using the Amplitude transformation package, include the following amplitude_source package version in your packages.yml
file.
TIP: Check dbt Hub for the latest installation instructions or read the dbt docs for more information on installing packages.
packages:
- package: fivetran/amplitude_source
version: [">=0.3.0", "<0.4.0"] # we recommend using ranges to capture non-breaking changes automatically
Step 3: Define database and schema variables
By default, this package runs using your destination and the amplitude_source
schema. If this is not where your Amplitude data is (for example, if your Amplitude schema is named amplitude_fivetran
), you would add the following configuration to your root dbt_project.yml
file with your custom database and schema names:
vars:
amplitude_database: your_destination_name
amplitude_schema: your_schema_name
Step 4: Configure event date range
Because of the typical volume of event data, you may want to limit this package's models to work with a more recent date range. Therefore we have added the functionality to filter records via the amplitude__date_range_start
and amplitude__date_range_end
variables within the stg_amplitude__event
model. The default date range starts at '2020-01-01' and ends one month past the current day, but you may configure it in your root dbt_project.yml
file:
vars:
amplitude__date_range_start: '2022-01-01' # your start date here
amplitude__date_range_end: '2022-12-01' # your end date here
If you adjust the date range variables, we recommend running dbt run --full-refresh
to ensure no data quality issues within the adjusted date range.
(Optional) Step 5: Additional configurations
Expand to view details
Change source table references
If an individual source table has a different name than the package expects, add the table name as it appears in your destination to the respective variable:
IMPORTANT: See this project's
dbt_project.yml
variable declarations to see the expected names.
vars:
amplitude_<default_source_table_name>_identifier: your_table_name
Change build schema
By default, this package builds the Amplitude staging models within a schema titled (<target_schema>
+ _source_amplitude
) in your destination. If this is not where you would like your Amplitude staging data to be written to, add the following configuration to your root dbt_project.yml
file:
models:
amplitude_source:
+schema: my_new_schema_name # leave blank for just the target_schema
(Optional) Step 6: Orchestrate your models with Fivetran Transformations for dbt Core™
Expand to view details
Fivetran offers the ability for you to orchestrate your dbt project through Fivetran Transformations for dbt Core™. Learn how to set up your project for orchestration through Fivetran in our Transformations for dbt Core™ setup guides.
Does this package have dependencies?
This dbt package is dependent on the following dbt packages. These dependencies are installed by default within this package. For more information on the following packages, refer to the dbt hub site.
IMPORTANT: If you have any of these dependent packages in your own
packages.yml
file, we highly recommend that you remove them from your rootpackages.yml
to avoid package version conflicts.
packages:
- package: fivetran/fivetran_utils
version: [">=0.4.0", "<0.5.0"]
- package: dbt-labs/dbt_utils
version: [">=1.0.0", "<2.0.0"]
- package: dbt-labs/spark_utils
version: [">=0.3.0", "<0.4.0"]
How is this package maintained and can I contribute?
Package Maintenance
The Fivetran team maintaining this package only maintains the latest version of the package. We highly recommend that you stay consistent with the latest version of the package and refer to the CHANGELOG and release notes for more information on changes across versions.
Contributions
A small team of analytics engineers at Fivetran develops these dbt packages. However, the packages are made better by community contributions.
We highly encourage and welcome contributions to this package. Check out this dbt Discourse article to learn how to contribute to a dbt package.
Are there any resources available?
- If you have questions or want to reach out for help, see the GitHub Issue section to find the right avenue of support for you.
- If you would like to provide feedback to the dbt package team at Fivetran or would like to request a new dbt package, fill out our Feedback Form.