Data Stream Quality Control with Apache Kafka
-
Updated
May 17, 2020 - Python
Data Stream Quality Control with Apache Kafka
Azure Data Lake Gen2 storage connectors for Data Culpa - monitor data quality automatically with Data Culpa Validator
MongoDB connector for Data Culpa - monitor data quality automatically with Data Culpa Validator
Snowflake connectors for Data Culpa - monitor data quality automatically with Data Culpa Validator
Generating Airflow DAG running soda data quality tests.
Open source clients for working with Data Culpa Validator services from data pipelines
Quality Aware Feature Store
Udacity's Data Engineering Nanodegree project: Data Pipeline with Airflow.
🕵️♀️ Enrich Metadata Features with AI Insights through Data Inspection
dbt Datasphere Plugin is for integrating multiple open-source data quality frameworks into your dbt projects. It unifies Soda SQL, Great Expectations, Datafold, providing a single interface to configure and run data quality checks.
(Meta)data Quality materials for Data Science Summer School 20023, Göttingen
Run greatexpectations.io on ANY SQL Engine using REST API. Supported by FastAPI, Pydantic and SQLAlchemy as best data quality tool
Data quality validations over PySpark DataFrame
⚡ Prevent downstream data quality issues by integrating the Soda Library into your CI/CD pipeline.
Find data quality issues and clean your data in a single line of code with a Scikit-Learn compatible Transformer.
Make your dataset talk to you. The AI assistant for data preparation.
data and pipeline testing with and for SQL
Add a description, image, and links to the dataquality topic page so that developers can more easily learn about it.
To associate your repository with the dataquality topic, visit your repo's landing page and select "manage topics."