Skip to content

Learn how to load data, create data models, add data quality tests and documentation using dbt Core with DuckDB

License

Notifications You must be signed in to change notification settings

pyladiesams/data-pipelines-with-dbtcore-duckdb-sep2025

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Building data pipelines with dbt Core and DuckDB

Workshop description

This hands-on workshop will introduce you to the fundamentals of dbt Core for building data pipelines. You will learn how to load data, create data models, add data quality tests and documentation using dbt Core with DuckDB as your database.

Requirements

  • PyLadies Amsterdam uses uv for dependency management
  • Google account if you want to use Google Colab

Usage

with uv

Run the following code:

git clone <github-url-of-workshop-repo>
cd <name-of-repo>

# create and activate venv, install dependencies
uv sync

with Google Colab

  1. Visit Google Colab
  2. In the top left corner select "File" → "Open Notebook"
  3. Under "GitHub", enter the URL of the repo of this workshop
  4. Select one of the notebooks within the repo.
  5. At the top of the notebook, add a Code cell and run the following code:
!git clone <github-url-of-workshop-repo>
%cd <name-of-repo>
!pip install -r requirements.txt

for a workshop giver

To get started, open the pyproject.toml file and set the required Python version. The pre-selected version 3.8 is generally a safe choice for most use cases.

After you have specified the Python version, you can create a virtual environment with uv venv and add packages with uv add <package>. Before the workshop, you can generate a requirements.txt file, which is needed, e.g. for running code in Google Colab, by running uv export > requirements.txt.

Video record

Re-watch this YouTube stream

Credits

This workshop was set up by @pyladiesams and @anyalitica

Appendix

Pre-Commit Hooks

To ensure our code looks beautiful, PyLadies uses pre-commit hooks. You can enable them by running pre-commit install. You may have to install pre-commit first, using uv sync, uv pip install pre-commit or pip install pre-commit.

Happy Coding :)

About

Learn how to load data, create data models, add data quality tests and documentation using dbt Core with DuckDB

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published