Personal project designed to improve my skills in data engineering.
It focuses on developing data pipelines that extract, transform, and load data into data warehouses. Additionally, it involves creating a dashboard with visualizations in Streamlit.
In this product, live worldwide city bike data is displayed on an interactive map with city bike details such as location, company, and city bike name.
Click this link to view the app
Check out changelog for older, current, and future updates.
Prefect | Streamlit | Docker | Google Cloud
BigQuery
(Orchastrated with Prefect)
- Data is extracted from City Bike API using Rapid API with Python.
- Data is moved from Python to BigQuery Database.
- BigQuery SQL Queries clean and transform data.
- Data is migrated for visualizations on Streamlit.
Github Actions
CodeQL is used to scan the repository for vulnerabilities in the Python packages.
Building, pushing, and deploying the Streamlit Docker image takes place across three jobs:
- Build the image
- Push the image to Artifact Registry
- Deploy the image to Cloud Run
Future plans include drawing more data from different sources and styles, more advanced CI/CD processes, and implementing more security. Check changelog for more info.