A data pipeline for Extracting, Loading and Transforming traffic flow data
A fully dockerized ELT pipeline using PostgreSQL, dbt, Apache Airflow, and Redash.
Explore the docs »
Table of Contents
A completely dockerized ELT pipeline with PostgreSQL for data storage, Airflow for automation and orchestration, DBT for data transformation, and a Redash dashboard connected to the data warehouse.
Tech Stack used in this project
Make sure you have docker installed on local machine.
- Docker
- Docker Compose
-
Clone the repo
git clone https://github.com/nahomfix/traffic-flow-ELT.git
-
Navigate to the folder
cd traffic-flow-ELT
-
Build an airflow image
docker build . --tag extending_airflow:latest
-
Run the following command once for first time initialization
docker-compose up airflow-init
-
Run
docker-compose up
-
Open Airflow web browser
Navigate to `http://localhost:8000/` on the browser activate and trigger load_dag activate and trigger transform_dag
-
Access redash dashboard
Navigate to `http://localhost:5000/` on the browser
Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.
- Fork the Project
- Create your Feature Branch (
git checkout -b feature/AmazingFeature
) - Commit your Changes (
git commit -m 'Add some AmazingFeature'
) - Push to the Branch (
git push origin feature/AmazingFeature
) - Open a Pull Request
Distributed under the MIT License. See LICENSE
for more information.
Nahom Bekele - nahom.fix@gmail.com