As a seasoned Business Intelligence Engineer, I bring over 4+ years of expertise in delivering data-driven solutions that propel business growth and optimize operations. My passion lies in building scalable data pipelines that leverage the latest tools and technologies to extract insights and drive decision-making.
I have a strong background in Python and the Hadoop Ecosystem, which enables me to develop efficient data pipelines that handle complex datasets with ease. I specialize in exploratory data analysis, structured machine learning algorithms, and visualization using Tableau & PowerBI, which ensures that I can extract valuable insights from data and present them in an easy-to-understand format.
My experience in data collection, cleaning, integration, processing, and visualization using libraries such as PySpark, Pandas, NumPy, and Shell makes me well-equipped to handle any data-related challenge. I have a solid understanding of ELT and ETL principles, and am proficient in SQL, AWS Redshift, and Snowflake with a strong background in data warehousing concepts. I also possess advanced skills in Excel and working knowledge in JIRA and Linux.
Apart from my technical skills, I possess advanced Excel skills and working knowledge of JIRA and Linux. With a proven track record of delivering data-driven solutions, I am confident in my ability to contribute to your organization's success.
Thank you for taking the time to read my profile, and I look forward to discussing how I can help your organization leverage data to drive success.
- Programming Languages: Python, Java
- Cloud: AWS, Azure, Databricks
- Databases: AWS RedShift, MySQL, PostgreSQL, MongoDB
- Data Integration / ETL: Informatica, AWS Glue
- Azure: Data Factory, Azure DB for MySQL, Blob Storages, Synapse Analytics
- AWS: Redshift, EC2, Quicksight, Kinesis, Sagemaker, Lambda Functions, S3, SNS
- Data Engineering: AirFlow (Orchestration), SnowFlake (Compute), Jenkins (CI/CD), Kafka (Streaming), Docker (Containers)
- BI Visualization: Tableau, Power BI, Spotfire
- Big Data: PySpark, Hadoop (HDFS, HIVE)
- IDE Tools: PyCharm, Eclipse, Jupyter Notebook, Google Colab, PyUnit (Testing Framework)
- Web Technologies: HTML, CSS, JavaScript, Angular, Flask, SpringBoot, Node.js
- Version Control: Git, BitBucket, Confluence Pages
- Data & Statistics: A/B testing, ETL, Data (Cleaning, Wrangling, Modeling, Visualization)
As a data engineer, I am always looking for ways to improve my craft and stay current with the latest trends and technologies. I welcome valuable suggestions and recommendations from others in the field, as well as from experts in related areas. I believe that by actively seeking out input and advice, I can continue to grow as a professional and deliver increasingly effective data engineering solutions. Whether it's through in-person discussions, online forums, or collaboration with others, I am always eager to learn and incorporate new ideas and best practices into my work.
Connect with me on Linkedin