Azure Databricks Notebook that assigs team members to customers based on a set of criteria
-
Updated
Jul 1, 2018 - Python
Azure Databricks Notebook that assigs team members to customers based on a set of criteria
Real World Project on Formula1 Racing using Azure Databricks, Delta Lake, Unity Catalog, Azure Data Factory [DP203]
Databricks provides a unified, open platform for all your data. It empowers data scientists, data engineers and data analysts with a simple collaborative environment to run interactive and scheduled data analysis workloads.
Code accelerator to migrate data from Snowflake tables into Databricks Delta Live Tables
Introducing Delta-Buddy: Your ultimate Delta Lake companion! 🚀 Streamline your data journey with an AI-powered chatbot. Ask Delta-Buddy anything about your Delta Lake.
Códigos em spark utilizados no dia a dia para manipulação de dados desde a ingestão até o refinamento.
Example of what you can do in Databricks in the Secure Data Environment (SDE) using Python, SQL, and R.
An Formula One ( F1 ) ETL project in Azure – Azure Data Lake - Data Factory - Azure Databricks data-engineering-project
Utlised Azure services to orchestrate earthquake daily data pipeline from USGS, consisting of last 7 days of data. Data was transformed and loaded into Azure SQL database and finally Tableau dashboard was made.
This project explores data analysis of the Indian Premier League utilizing AWS S3, Apache Spark, python, and SQL.
The project harnessed an ETL multi-hop architecture, ingesting data from the Ergast API into a storage backed by Azure Data Lake. The process involved weekly ingestion of bronze layer data as cutover and delta files. Raw data, in varied formats, was transformed using Azure Databricks PySpark notebooks into enriched Silver and Gold layers.
Jupybricks is a python package that allows databricks developers to switch easily between local development in jupyter notebooks and databricks notebooks. 🔨 convert databricks .py files into jupyter notebooks . 🔨 convert local jupyter notebooks to databricks .py files
Accelerator code for advertising spend optimization, leveraging Databricks
Repository to develop databricks notebooks suitable for use in a Azure DevOps environment, and space for refactoring existing ETL design
Accelerator code for an anomaly detection module leveraging Databricks for use as part of a Network Threat Detection System
Big data processing of news with Text Mining in Apache Spark through 3 fundamental processes: data preparation, searching based on the inverted index and grouping of news by similarity.
An introduction to PySpark, Creating a simple multi regression ML model and hosting it on a databricks cluster
A bilingual ChatBot created to answer to FAQs asked to the Colombian Institute for Family Welfare (ICBF).
nbmanips allows you easily manipulate ipynb files
Continuous Delivery tool for PySpark Notebooks based jobs on Databricks
Add a description, image, and links to the databricks-notebooks topic page so that developers can more easily learn about it.
To associate your repository with the databricks-notebooks topic, visit your repo's landing page and select "manage topics."