Skip to content

Example of implementing data pipelines for ingesting 3rd party data into a Snowflake database

License

Notifications You must be signed in to change notification settings

davorborcic/data_pipelines

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Creating Data Pipelines in Airflow

Example of implementing data pipelines for ingesting 3rd party data into a Snowflake database

The code corresponds to my article Creating Data Pipelines in Airflow The best way to understand the code and its purpose is to go through the article.

Prerequisites

Instructions

  1. Clone the repository
  2. Download the NBA file and save it as nba_elo.csv in files\staging subfolder
  3. Connect to the Snowflake and run create_database_objects.sql

Start the Docker, configuring Airflow, and initiating the data pipeline

  1. From data_pipelines folder execute docker compose up
  2. Log in to https://localhost:8080 (airflow / airflow)
  3. Create Snowflake connection named sf1
  4. Enable DAG
  5. Move or copy the nba_elo.csv from staging subfolder into files folder
  6. Observe the data pipeline running and loading the data into the database

Credits

About

Example of implementing data pipelines for ingesting 3rd party data into a Snowflake database

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages