Skip to content
View James-Wachuka's full-sized avatar
  • Kenya
Block or Report

Block or report James-Wachuka

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
James-Wachuka/README.md

About me

I am a passionate and skilled data engineer πŸ‘¨β€πŸ’» with three years of experience in the field. My expertise lies in using tools like GCP, Python, Spark, SQL, Talend Open Studio, Tableau, Power BI, Docker, Terraform , dbt, and Airflow . I have successfully designed and developed robust data pipelines and architectures, ensuring data integrity and accuracy βœ”οΈ. With my expertise in cloud computing platforms like GCP ☁️, I have implemented scalable solutions. I love transforming complex data sets into actionable insights using visualization tools like Tableau and Power BI πŸ“ˆ. I have also mastered containerization πŸ“¦, infrastructure provisioning βš™οΈ, and workflow orchestration 🎼. I am driven to deliver excellence in data engineering and thrive on pushing boundaries πŸ’ͺ. Request resume.

I also write about data engineering on

🌐communities:@DataTalksClub @TechUp Africa

πŸ“« lets talk: mail me at jewachu26@gmail.com or through my socials

𝕏.com linkedin spotify

Skills


β€’ Data Modelling β€’ DataOps β€’ Data versioning and source control β€’ Cloud data engineering β€’ Extract Transform Load (ETL) β€’ Data Warehousing β€’ Data stream and real time processing β€’ Orchestration & Workflow automation β€’ Data governance and quality management

Tech stack




Latest Articles

TitleLink
Running Transformations on BigQuery using dbt Cloud: step by stephttps://dev.to/wachuka_james/running-transformations-on-bigquery-using-dbt-cloud-step-by-step-11bo
Debugging Python Data Pipelineshttps://dev.to/wachuka_james/debugging-python-data-pipelines-a-step-by-step-guide-11g7
Using pyspark to stream data from coingecko API and visualise using dashhttps://dev.to/wachuka_james/using-pyspark-to-stream-data-from-coingecko-api-and-visualise-using-dash-5g43

Pinned

  1. mentalhealth_analysis-data-pipeline mentalhealth_analysis-data-pipeline Public

    An end to end data pipeline for for mental health analysis

    Python

  2. podcasts_pipeline podcasts_pipeline Public

    Building a four-step data pipeline using Airflow to download podcast episodes.

    Python

  3. Prefect-PostgreSQL-Sensors Prefect-PostgreSQL-Sensors Public

    The prefect_postgres_sensors package provides Prefect sensors for monitoring changes or conditions within a PostgreSQL database.

    Python

  4. event-driven-microservices event-driven-microservices Public

    This project demonstrates an event-driven microservices architecture using Apache Kafka for event streaming and webhook integration with external services

    Python 1

  5. python-kafka_distributed_task_queue python-kafka_distributed_task_queue Public

    a simple implementation of a distributed task queue

    Python 1