UC Davis Distributed Computing with Spark SQL (with Databricks) and Databricks Apache Spark SQL for Data Analysts
-
Updated
Jul 10, 2021 - HTML
UC Davis Distributed Computing with Spark SQL (with Databricks) and Databricks Apache Spark SQL for Data Analysts
Apply Data Engineering to Personal Finance
Collect & stream data, notify user on Telegram
As a Data Engineer for a fictional E-commerce startup, this project addresses the task of analyzing the web server logs to find the number of product pages visited and the number of items in the cart.
A comprehensive repository housing a collection of insightful blog posts, in-depth documentation, and resources exploring various facets of data engineering. From ETL processes and database management to orchestration tools, data quality, monitoring, and deployment strategies
Developed a deep learning model utilizing TensorFlow to automate the classification of financial documents. Leveraging a Bidirectional LSTM RNN, we accurately categorize the documents. Our user-friendly Streamlit application ensures high accuracy & efficiency in document management, all deployed on the Hugging Face platform for seamless integration
Neste projeto de Análise de Recursos Humanos, temos como objetivo responder questões-chave sobre gestão de talentos e rotatividade de colaboradores em uma empresa fictícia.
Bringing you the posts that matter.
Speed comparisons for dataframe libraries
A data-pipeline for high-resolution power meter data
This is the sentiment analysis on the #VisitRwanda on twitter, this is a campaign of eco tourism in Rwanda that promotes touristic places attraction in Rwanda.
Data science and Spark applied to 7 hypotheses regarding the DJIA stock ticker and daily news.
A data-pipeline for high-resolution power meter data
Automated Tool for Optimized Modelling
Streaming data changes to a Data Lake with Debezium and Delta Lake pipeline
The dbt-native data observability solution for data & analytics engineers. Monitor your data pipelines in minutes. Available as self-hosted or cloud service with premium features.
Add a description, image, and links to the data-pipeline topic page so that developers can more easily learn about it.
To associate your repository with the data-pipeline topic, visit your repo's landing page and select "manage topics."