GithubHelp home page GithubHelp logo

fivetran / dbt_pardot Goto Github PK

View Code? Open in Web Editor NEW
1.0 39.0 3.0 791 KB

Fivetran's Pardot dbt package

Home Page: https://fivetran.github.io/dbt_pardot/

License: Apache License 2.0

Shell 100.00%
pardot dbt dbt-packages fivetran

dbt_pardot's Introduction

Pardot Transformation dbt Package (Docs)

📣 What does this dbt package do?

  • Produces modeled tables that beverage Pardot data from Fivetran's connector in the format described by this ERD and builds off the output of our Pardot source package.
  • Enables you to better understand your Pardot prospects, opportunities, lists, and campaign performance.
  • Generates a comprehensive data dictionary of your source and modeled Pardot data through the dbt docs site.

The following table provides a detailed list of all models materialized within this package by default.

TIP: See more details about these models in the package's dbt docs site. .

Models

This package contains transformation models, designed to work simultaneously with our Pardot source package. A dependency on the source package is declared in this package's packages.yml file, so it will automatically download when you run dbt deps. The primary outputs of this package are described below. Intermediate models are used to create these output models.

Model Description
pardot__campaigns Each record represents a campaign in Pardot, enriched with metrics about associated prospects.
pardot__lists Each record represents a list in Pardot, enriched with metrics about associated prospect activity.
pardot__opportunities Each record represents an opportunity in Pardot, enriched with metrics about associated prospects.
pardot__prospects Each record represents a prospect in Pardot, enriched with metrics about associated prospect activity.

🎯 How do I use the dbt package?

Step 1: Prerequisites

To use this dbt package, you must have the following:

  • At least one Fivetran Pardot connector syncing data into your destination.
  • A BigQuery, Snowflake, Redshift, or PostgreSQL destination.

Step 2: Install the package

Include the following pardot package version in your packages.yml file:

TIP: Check dbt Hub for the latest installation instructions or read the dbt docs for more information on installing packages.

packages:
  - package: fivetran/pardot
    version: [">=0.5.0", "<0.6.0"]

Do NOT include the pardot_source package in this file. The transformation package itself has a dependency on it and will install the source package as well.

Step 3: Define database and schema variables

By default, this package runs using your destination and the pardot schema. If this is not where your Pardot data is (for example, if your Pardot schema is named pardot_fivetran), add the following configuration to your root dbt_project.yml file:

```yml
vars:
  pardot_source:
    pardot_database: your_database_name
    pardot_schema: your_schema_name 

(Optional) Step 4: Additional configurations

Expand for configurations

Passthrough Columns

By default, the package includes all of the standard columns in the stg_pardot__prospect model. If you want to include custom columns, configure them using the prospect_passthrough_columns variable:

vars:
  pardot_source:
    prospect_passthrough_columns: ["custom_creative","custom_contact_state"]

Additional metrics

By default, this package aggregates and joins activity data onto the prospect model for email and visit events. If you want to have aggregates for other events in the visitor_activity table, use prospect_metrics_activity_types variable to generate these aggregates. Use the type_name column value:

vars:
  pardot:
    prospect_metrics_activity_types: ["form handler","webinar"]  

Changing the Build Schema

By default this package will build the Pardot staging models within a schema titled (<target_schema> + _stg_pardot) and Pardot final models within a schema titled (<target_schema> + pardot) in your target database. If this is not where you would like your modeled Pardot data to be written, add the following configuration to your dbt_project.yml file:

models:
    pardot:
      +schema: my_new_schema_name # leave blank for just the target_schema
    pardot_source:
      +schema: my_new_schema_name # leave blank for just the target_schema

Change the source table references

If an individual source table has a different name than the package expects, add the table name as it appears in your destination to the respective variable:

IMPORTANT: See this project's dbt_project.yml variable declarations to see the expected names.

vars:
    pardot_<default_source_table_name>_identifier: your_table_name 

(Optional) Step 6: Orchestrate your models with Fivetran Transformations for dbt Core™

Expand to view details

Fivetran offers the ability for you to orchestrate your dbt project through Fivetran Transformations for dbt Core™. Learn how to set up your project for orchestration through Fivetran in our Transformations for dbt Core™ setup guides.

🔍 Does this package have dependencies?

This dbt package is dependent on the following dbt packages. Please be aware that these dependencies are installed by default within this package. For more information on the following packages, refer to the dbt hub site.

IMPORTANT: If you have any of these dependent packages in your own packages.yml file, we highly recommend that you remove them from your root packages.yml to avoid package version conflicts.

packages:
    - package: fivetran/fivetran_utils
      version: [">=0.4.0", "<0.5.0"]

    - package: dbt-labs/dbt_utils
      version: [">=1.0.0", "<2.0.0"]

    - package: fivetran/pardot_source
      version: [">=0.5.0", "<0.6.0"]

🙌 How is this package maintained and can I contribute?

Package Maintenance

The Fivetran team maintaining this package only maintains the latest version of the package. We highly recommend you stay consistent with the latest version of the package and refer to the CHANGELOG and release notes for more information on changes across versions.

Contributions

A small team of analytics engineers at Fivetran develops these dbt packages. However, the packages are made better by community contributions!

We highly encourage and welcome contributions to this package. Check out this dbt Discourse article on the best workflow for contributing to a package!

🏪 Are there any resources available?

  • If you have questions or want to reach out for help, please refer to the GitHub Issue section to find the right avenue of support for you.
  • If you would like to provide feedback to the dbt package team at Fivetran or would like to request a new dbt package, fill out our Feedback Form.
  • Have questions or want to just say hi? Book a time during our office hours on Calendly or email us at [email protected].

dbt_pardot's People

Contributors

dylanbaker avatar fivetran-catfritz avatar fivetran-chloe avatar fivetran-jamie avatar fivetran-joemarkiewicz avatar fivetran-kristin avatar fivetran-reneeli avatar fivetran-sheringuyen avatar jlmendgom5tran avatar

Stargazers

 avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

dbt_pardot's Issues

[Feature] Databricks Compatibility

Copied from fivetran/dbt_mixpanel #34.

Is there an existing feature request for this?

  • I have searched the existing issues

Describe the Feature

For Databricks Compatibility, add the following:

  1. Buildkite testing:
    • Update pre-command (example)
    • Update pipeline.yml (example)
    • Update sample.profiles.yml (example)
    • Add the below to integration_tests/dbt_project.yml if it's not there:
dispatch:
  - macro_namespace: dbt_utils
    search_order: ['spark_utils', 'dbt_utils']
  1. For source packages, update src yml so a database won't be passed to spark (example or use below):
sources: 
  - name: <name>
    database: "{% if target.type != 'spark' %}{{ var('<name>_database', target.database) }}{% endif %}"
  1. Update any incremental models to update partition_by for databricks and add current strategies if not present:
config(
        materialized='incremental',
        unique_key='<original unique key>',
        partition_by={'field': '<original field>', 'data_type': '<original data type>'} if target.type not in ('spark','databricks') else ['<original field>'],
        incremental_strategy = 'merge' if target.type not in ('postgres', 'redshift') else 'delete+insert',
        file_format = 'delta' 
)

Describe alternatives you've considered

No response

Are you interested in contributing this feature?

  • Yes.
  • Yes, but I will need assistance and will schedule time during your office hours for guidance.
  • No.

Anything else?

No response

[Feature] Update README

Is there an existing feature request for this?

  • I have searched the existing issues

Describe the Feature

The README needs to updated to the current format.

Describe alternatives you've considered

No response

Are you interested in contributing this feature?

  • Yes.
  • Yes, but I will need assistance and will schedule time during your office hours for guidance.
  • No.

Anything else?

No response

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.