-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GE errors when trying to return a result when contains a pandas NA #2029
Comments
@isichei Thank you for reporting this! If you would like to make a PR (the fix must be very small), it would be super welcome! Please let us know. |
Yeah can try and have a look over the next couple of weeks. When you say If it is a bigger fix would you want me to make several small PRs? |
@isichei I meant "must" not as in "we require the fix to be small", but as in "I think it will not take much code to fix this" :) Looking forward to your PR! |
Please enter the commit message for your changes. Lines starting
Signed-off-by: James Campbell <james.p.campbell@gmail.com>
* Fix/deprecate test_column_reflection_fallback * requested_tests * Add tests for splitters and tests for samplers. * Added a comment. * Comment. * Managing PartitionQuery properly. * Update great_expectations/execution_environment/data_connector/data_connector.py Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * making sure this runs first * Update stored meta batch_spec and batch_markers Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Refactoring splitting and sampling tests, partially out of the Golden Path. * Re-enable connection for sqlite, snowflake, mssql Signed-off-by: James Campbell <james.p.campbell@gmail.com> * adding spark fixture * Running Isort to fix lint errors. * linting * isort * Linting. * big query temporary table name * linting * WIP Core Concepts Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * lint * lint * cleanup * Add docs gitignore * Typo. * Tidy and use SparkDFBatchData * Bring changes from #2029 / #2039 Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix merge errors. * Lint. * comment * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Restoring BatchKwargs for the Legacy backward compatibility. * Linting * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Switch SparkDFBatchData into try-except import block * Lint * Update how_to_configure_a_redshift_datasource.rst * small formatting change * Additional decoupling from Legacy datasource implementation. * Typo * Linting. * Anonymizers * Anonymizers * Anonymizers * Linting * add_column_row_condition fully tested * Initial tests for execution engine parent class passing * Execution Engine tests finished * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update doc * Rename creating_modular_expectations.rst to how_to_create_modular_expectations.rst * Turn code into code blocks * Significant updates to formatting Almost ready for review * More formatting additions * Ready for review * WIP * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * last bit of clean up. how-to-redshift-datasource * update how-to-write-how-to and clean up * typo * Guide finished * updates from review * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * first push before PR * Update docs/conf.py * Update how_to_configure_a_pandas_filesystem_datasource.rst * Corrections such as unneeded method and needed Anonymizer. * remove legacy diff description * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * first push for how-to-sparkdf-filesystem doc * added some more references * first push. Before adding blurb on introspection and query * first push of doc + bugfix for query * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * added output * Formatting issues but information much improved * Update docs/conf.py * Slight formatting improvement * first push how-to-mysql-datasource * Complete for review pt.2 * some formatting changes * Update docs/conf.py * Renaming Datasource to LegacyDatasource, DatasourceConfig to LegacyDatasourceConfig, DatasourceConfigSchema to LegacyDatasourceConfigSchema, and datasourceConfigSchema to legacyDatasourceConfigSchema. * s/StreamlinedSqlExecutionEngine/SimpleSqlalchemyDatasource/ * Renaming ExecutionEnvironment to Datasource * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Linting. * renaming execution_environment into new_datasource WIP * renaming execution_environment into new_datasource WIP * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * removed spark output from doc * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_s3_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * updates from Pr review (how to pandas s3) * trailing line * renamed test_asset * ExecutionEnvironment -> Datasource * Linting. * oops * dotting some i's * Merge. * Clean up. * Clean up. * Clean up. * Linting. * updates from a closer look * String literals renaming. * Remove unusued execution_environment code from base.py and add a comment to make Schema validation robust for the new and the Legacy Datasource classes alike. * Update how_to_configure_a_snowflake_datasource.rst * Add sample pngs * Add doc for How to Create Renderers for Custom Expectations * Update renderer to return typed RenderedTableContent * Add typehints * Add sample images * Add page ref * Fill out how-to guide * Crop image * lint * Remove batch_definition from get_batch and get_validator * Tidy up error handling * Add much better tests for get_validator; Switch attach_new_expectation_suite to create_expectation_suite_with_name * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Propagate name changes through new tests * Make black happy * Apply suggestions from code review removed `role: ADMIN` and `warehouse` from config * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Make isort happy * Add docstrings for MetricProvider and Expectation Signed-off-by: James Campbell <james.p.campbell@gmail.com> * PR Review updates Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete .test_durations * Delete Untitled.ipynb * Delete ge_docs_links.csv * Commit cleanup Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Ensure BatchDefinition is serializable Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix import and custom expectation issues Signed-off-by: James Campbell <james.p.campbell@gmail.com> * WIP custom expectations docs fixes Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Use legacy logic for default expectation values. Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * Update link Signed-off-by: James Campbell <james.p.campbell@gmail.com> * 1. how_to_configure_a_redshift_datasource: a. Added step 5 - save the config - and modified the note in step 4 b. Fixed the class name - it was out of date 2. how_to_configure_a_snowflake_datasource: Fixed the class name in the output snippet - it was out of date * Update changelog, version for release Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete tmp--pyproject.toml * Spark Self-Managed WIP * Self-managed Spark WIP * Self-managed Spark documentation/UAT WIP. * Self-managed Spark documentation/UAT WIP. * reset * Pin to legacy pip Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Remove TableDataConnector -- it is a broken unimplemented module that was copied and pasted from Legacy design, but not worked on. * Self-managed Spark documentation/UAT WIP. * UAT-based updates to parameterized expectations Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Docs/draft docs for data connectors (#2086) Changes proposed in this pull request: This PR adds how-to guides for new-style DataConnectors and Datasources for 0.13. Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: William Shin <will@superconductive.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update links * Add docs tab for experimental API * Add admonition pointing user to docs for experimental API * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Abe <abegong@gmail.com> Co-authored-by: William Shin <will@superconductive.com> Co-authored-by: Alex Sherstinsky <alex@superconductivehealth.com> Co-authored-by: Alex Sherstinsky <alexsherstinsky@users.noreply.github.com> Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Rob Lim <robert.m.lim@gmail.com> Co-authored-by: gilpasternak35 <gilpasternak35@gmail.com> Co-authored-by: Eugene Mandel <eugene@superconductivehealth.com>
* Re-enable connection for sqlite, snowflake, mssql Signed-off-by: James Campbell <james.p.campbell@gmail.com> * adding spark fixture * Running Isort to fix lint errors. * linting * isort * Linting. * big query temporary table name * linting * WIP Core Concepts Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * lint * lint * cleanup * Add docs gitignore * Typo. * Tidy and use SparkDFBatchData * Bring changes from #2029 / #2039 Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix merge errors. * Lint. * comment * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Restoring BatchKwargs for the Legacy backward compatibility. * Linting * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Switch SparkDFBatchData into try-except import block * Lint * Update how_to_configure_a_redshift_datasource.rst * small formatting change * Additional decoupling from Legacy datasource implementation. * Typo * Linting. * Anonymizers * Anonymizers * Anonymizers * Linting * add_column_row_condition fully tested * Initial tests for execution engine parent class passing * Execution Engine tests finished * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update doc * Rename creating_modular_expectations.rst to how_to_create_modular_expectations.rst * Turn code into code blocks * Significant updates to formatting Almost ready for review * More formatting additions * Ready for review * WIP * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * last bit of clean up. how-to-redshift-datasource * update how-to-write-how-to and clean up * typo * Guide finished * updates from review * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * first push before PR * Update docs/conf.py * Update how_to_configure_a_pandas_filesystem_datasource.rst * Corrections such as unneeded method and needed Anonymizer. * remove legacy diff description * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * first push for how-to-sparkdf-filesystem doc * added some more references * first push. Before adding blurb on introspection and query * first push of doc + bugfix for query * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * added output * Formatting issues but information much improved * Update docs/conf.py * Slight formatting improvement * first push how-to-mysql-datasource * Complete for review pt.2 * some formatting changes * Update docs/conf.py * Renaming Datasource to LegacyDatasource, DatasourceConfig to LegacyDatasourceConfig, DatasourceConfigSchema to LegacyDatasourceConfigSchema, and datasourceConfigSchema to legacyDatasourceConfigSchema. * s/StreamlinedSqlExecutionEngine/SimpleSqlalchemyDatasource/ * Renaming ExecutionEnvironment to Datasource * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Linting. * renaming execution_environment into new_datasource WIP * renaming execution_environment into new_datasource WIP * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * removed spark output from doc * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_s3_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * updates from Pr review (how to pandas s3) * trailing line * renamed test_asset * ExecutionEnvironment -> Datasource * Linting. * oops * dotting some i's * Merge. * Clean up. * Clean up. * Clean up. * Linting. * updates from a closer look * String literals renaming. * Remove unusued execution_environment code from base.py and add a comment to make Schema validation robust for the new and the Legacy Datasource classes alike. * Update how_to_configure_a_snowflake_datasource.rst * Add sample pngs * Add doc for How to Create Renderers for Custom Expectations * Update renderer to return typed RenderedTableContent * Add typehints * Add sample images * Add page ref * Fill out how-to guide * Crop image * lint * Remove batch_definition from get_batch and get_validator * Tidy up error handling * Add much better tests for get_validator; Switch attach_new_expectation_suite to create_expectation_suite_with_name * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Propagate name changes through new tests * Make black happy * Apply suggestions from code review removed `role: ADMIN` and `warehouse` from config * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Make isort happy * Add docstrings for MetricProvider and Expectation Signed-off-by: James Campbell <james.p.campbell@gmail.com> * PR Review updates Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete .test_durations * Delete Untitled.ipynb * Delete ge_docs_links.csv * Commit cleanup Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Ensure BatchDefinition is serializable Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix import and custom expectation issues Signed-off-by: James Campbell <james.p.campbell@gmail.com> * WIP custom expectations docs fixes Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Use legacy logic for default expectation values. Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * Update link Signed-off-by: James Campbell <james.p.campbell@gmail.com> * 1. how_to_configure_a_redshift_datasource: a. Added step 5 - save the config - and modified the note in step 4 b. Fixed the class name - it was out of date 2. how_to_configure_a_snowflake_datasource: Fixed the class name in the output snippet - it was out of date * Update changelog, version for release Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete tmp--pyproject.toml * Spark Self-Managed WIP * Self-managed Spark WIP * Self-managed Spark documentation/UAT WIP. * Self-managed Spark documentation/UAT WIP. * reset * Pin to legacy pip Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Remove TableDataConnector -- it is a broken unimplemented module that was copied and pasted from Legacy design, but not worked on. * Self-managed Spark documentation/UAT WIP. * UAT-based updates to parameterized expectations Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Docs/draft docs for data connectors (#2086) Changes proposed in this pull request: This PR adds how-to guides for new-style DataConnectors and Datasources for 0.13. Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: William Shin <will@superconductive.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Enhance self_check of Datasource class. * Linting * ExecutionEngine Config [WIP] * ExecutionEngine Config [WIP] * Update links * Add docs tab for experimental API * Add admonition pointing user to docs for experimental API * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * ExecutionEngine Config * Tests for more expressive Datasource.self_check() diagnostic. * Linting. * TypeError fix for numeric utility. * Additional test. Co-authored-by: James Campbell <james.p.campbell@gmail.com> Co-authored-by: Rob Lim <robert.m.lim@gmail.com> Co-authored-by: William Shin <will@superconductive.com> Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Abe <abegong@gmail.com> Co-authored-by: gilpasternak35 <gilpasternak35@gmail.com> Co-authored-by: Eugene Mandel <eugene@superconductivehealth.com>
Please enter the commit message for your changes. Lines starting Co-authored-by: Eugene Mandel <eugene.mandel@gmail.com>
* Fix/deprecate test_column_reflection_fallback * requested_tests * Add tests for splitters and tests for samplers. * Added a comment. * Comment. * Managing PartitionQuery properly. * Update great_expectations/execution_environment/data_connector/data_connector.py Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * making sure this runs first * Update stored meta batch_spec and batch_markers Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Refactoring splitting and sampling tests, partially out of the Golden Path. * Re-enable connection for sqlite, snowflake, mssql Signed-off-by: James Campbell <james.p.campbell@gmail.com> * adding spark fixture * Running Isort to fix lint errors. * linting * isort * Linting. * big query temporary table name * linting * WIP Core Concepts Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * lint * lint * cleanup * Add docs gitignore * Typo. * Tidy and use SparkDFBatchData * Bring changes from great-expectations#2029 / great-expectations#2039 Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix merge errors. * Lint. * comment * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Restoring BatchKwargs for the Legacy backward compatibility. * Linting * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Switch SparkDFBatchData into try-except import block * Lint * Update how_to_configure_a_redshift_datasource.rst * small formatting change * Additional decoupling from Legacy datasource implementation. * Typo * Linting. * Anonymizers * Anonymizers * Anonymizers * Linting * add_column_row_condition fully tested * Initial tests for execution engine parent class passing * Execution Engine tests finished * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update doc * Rename creating_modular_expectations.rst to how_to_create_modular_expectations.rst * Turn code into code blocks * Significant updates to formatting Almost ready for review * More formatting additions * Ready for review * WIP * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * last bit of clean up. how-to-redshift-datasource * update how-to-write-how-to and clean up * typo * Guide finished * updates from review * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * first push before PR * Update docs/conf.py * Update how_to_configure_a_pandas_filesystem_datasource.rst * Corrections such as unneeded method and needed Anonymizer. * remove legacy diff description * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * first push for how-to-sparkdf-filesystem doc * added some more references * first push. Before adding blurb on introspection and query * first push of doc + bugfix for query * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * added output * Formatting issues but information much improved * Update docs/conf.py * Slight formatting improvement * first push how-to-mysql-datasource * Complete for review pt.2 * some formatting changes * Update docs/conf.py * Renaming Datasource to LegacyDatasource, DatasourceConfig to LegacyDatasourceConfig, DatasourceConfigSchema to LegacyDatasourceConfigSchema, and datasourceConfigSchema to legacyDatasourceConfigSchema. * s/StreamlinedSqlExecutionEngine/SimpleSqlalchemyDatasource/ * Renaming ExecutionEnvironment to Datasource * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Linting. * renaming execution_environment into new_datasource WIP * renaming execution_environment into new_datasource WIP * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * removed spark output from doc * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_s3_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * updates from Pr review (how to pandas s3) * trailing line * renamed test_asset * ExecutionEnvironment -> Datasource * Linting. * oops * dotting some i's * Merge. * Clean up. * Clean up. * Clean up. * Linting. * updates from a closer look * String literals renaming. * Remove unusued execution_environment code from base.py and add a comment to make Schema validation robust for the new and the Legacy Datasource classes alike. * Update how_to_configure_a_snowflake_datasource.rst * Add sample pngs * Add doc for How to Create Renderers for Custom Expectations * Update renderer to return typed RenderedTableContent * Add typehints * Add sample images * Add page ref * Fill out how-to guide * Crop image * lint * Remove batch_definition from get_batch and get_validator * Tidy up error handling * Add much better tests for get_validator; Switch attach_new_expectation_suite to create_expectation_suite_with_name * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Propagate name changes through new tests * Make black happy * Apply suggestions from code review removed `role: ADMIN` and `warehouse` from config * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Make isort happy * Add docstrings for MetricProvider and Expectation Signed-off-by: James Campbell <james.p.campbell@gmail.com> * PR Review updates Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete .test_durations * Delete Untitled.ipynb * Delete ge_docs_links.csv * Commit cleanup Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Ensure BatchDefinition is serializable Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix import and custom expectation issues Signed-off-by: James Campbell <james.p.campbell@gmail.com> * WIP custom expectations docs fixes Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Use legacy logic for default expectation values. Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * Update link Signed-off-by: James Campbell <james.p.campbell@gmail.com> * 1. how_to_configure_a_redshift_datasource: a. Added step 5 - save the config - and modified the note in step 4 b. Fixed the class name - it was out of date 2. how_to_configure_a_snowflake_datasource: Fixed the class name in the output snippet - it was out of date * Update changelog, version for release Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete tmp--pyproject.toml * Spark Self-Managed WIP * Self-managed Spark WIP * Self-managed Spark documentation/UAT WIP. * Self-managed Spark documentation/UAT WIP. * reset * Pin to legacy pip Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Remove TableDataConnector -- it is a broken unimplemented module that was copied and pasted from Legacy design, but not worked on. * Self-managed Spark documentation/UAT WIP. * UAT-based updates to parameterized expectations Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Docs/draft docs for data connectors (great-expectations#2086) Changes proposed in this pull request: This PR adds how-to guides for new-style DataConnectors and Datasources for 0.13. Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: William Shin <will@superconductive.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update links * Add docs tab for experimental API * Add admonition pointing user to docs for experimental API * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Abe <abegong@gmail.com> Co-authored-by: William Shin <will@superconductive.com> Co-authored-by: Alex Sherstinsky <alex@superconductivehealth.com> Co-authored-by: Alex Sherstinsky <alexsherstinsky@users.noreply.github.com> Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Rob Lim <robert.m.lim@gmail.com> Co-authored-by: gilpasternak35 <gilpasternak35@gmail.com> Co-authored-by: Eugene Mandel <eugene@superconductivehealth.com>
…xpectations#2091) * Re-enable connection for sqlite, snowflake, mssql Signed-off-by: James Campbell <james.p.campbell@gmail.com> * adding spark fixture * Running Isort to fix lint errors. * linting * isort * Linting. * big query temporary table name * linting * WIP Core Concepts Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * lint * lint * cleanup * Add docs gitignore * Typo. * Tidy and use SparkDFBatchData * Bring changes from great-expectations#2029 / great-expectations#2039 Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix merge errors. * Lint. * comment * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Restoring BatchKwargs for the Legacy backward compatibility. * Linting * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Moving the batch_kwargs module to its original (Legacy Implementation) location. * Switch SparkDFBatchData into try-except import block * Lint * Update how_to_configure_a_redshift_datasource.rst * small formatting change * Additional decoupling from Legacy datasource implementation. * Typo * Linting. * Anonymizers * Anonymizers * Anonymizers * Linting * add_column_row_condition fully tested * Initial tests for execution engine parent class passing * Execution Engine tests finished * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update doc * Rename creating_modular_expectations.rst to how_to_create_modular_expectations.rst * Turn code into code blocks * Significant updates to formatting Almost ready for review * More formatting additions * Ready for review * WIP * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_redshift_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * last bit of clean up. how-to-redshift-datasource * update how-to-write-how-to and clean up * typo * Guide finished * updates from review * Update how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * first push before PR * Update docs/conf.py * Update how_to_configure_a_pandas_filesystem_datasource.rst * Corrections such as unneeded method and needed Anonymizer. * remove legacy diff description * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * first push for how-to-sparkdf-filesystem doc * added some more references * first push. Before adding blurb on introspection and query * first push of doc + bugfix for query * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Core Concepts Update Signed-off-by: James Campbell <james.p.campbell@gmail.com> * added output * Formatting issues but information much improved * Update docs/conf.py * Slight formatting improvement * first push how-to-mysql-datasource * Complete for review pt.2 * some formatting changes * Update docs/conf.py * Renaming Datasource to LegacyDatasource, DatasourceConfig to LegacyDatasourceConfig, DatasourceConfigSchema to LegacyDatasourceConfigSchema, and datasourceConfigSchema to legacyDatasourceConfigSchema. * s/StreamlinedSqlExecutionEngine/SimpleSqlalchemyDatasource/ * Renaming ExecutionEnvironment to Datasource * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_modular_expectations.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Linting. * renaming execution_environment into new_datasource WIP * renaming execution_environment into new_datasource WIP * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_filesystem_datasource.rst * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_spark_filesystem_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * removed spark output from doc * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_pandas_s3_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * updates from Pr review (how to pandas s3) * trailing line * renamed test_asset * ExecutionEnvironment -> Datasource * Linting. * oops * dotting some i's * Merge. * Clean up. * Clean up. * Clean up. * Linting. * updates from a closer look * String literals renaming. * Remove unusued execution_environment code from base.py and add a comment to make Schema validation robust for the new and the Legacy Datasource classes alike. * Update how_to_configure_a_snowflake_datasource.rst * Add sample pngs * Add doc for How to Create Renderers for Custom Expectations * Update renderer to return typed RenderedTableContent * Add typehints * Add sample images * Add page ref * Fill out how-to guide * Crop image * lint * Remove batch_definition from get_batch and get_validator * Tidy up error handling * Add much better tests for get_validator; Switch attach_new_expectation_suite to create_expectation_suite_with_name * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Propagate name changes through new tests * Make black happy * Apply suggestions from code review removed `role: ADMIN` and `warehouse` from config * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_snowflake_datasource.rst * Make isort happy * Add docstrings for MetricProvider and Expectation Signed-off-by: James Campbell <james.p.campbell@gmail.com> * PR Review updates Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete .test_durations * Delete Untitled.ipynb * Delete ge_docs_links.csv * Commit cleanup Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Ensure BatchDefinition is serializable Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Fix import and custom expectation issues Signed-off-by: James Campbell <james.p.campbell@gmail.com> * WIP custom expectations docs fixes Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Use legacy logic for default expectation values. Signed-off-by: James Campbell <james.p.campbell@gmail.com> * lint * Update link Signed-off-by: James Campbell <james.p.campbell@gmail.com> * 1. how_to_configure_a_redshift_datasource: a. Added step 5 - save the config - and modified the note in step 4 b. Fixed the class name - it was out of date 2. how_to_configure_a_snowflake_datasource: Fixed the class name in the output snippet - it was out of date * Update changelog, version for release Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Delete tmp--pyproject.toml * Spark Self-Managed WIP * Self-managed Spark WIP * Self-managed Spark documentation/UAT WIP. * Self-managed Spark documentation/UAT WIP. * reset * Pin to legacy pip Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Remove TableDataConnector -- it is a broken unimplemented module that was copied and pasted from Legacy design, but not worked on. * Self-managed Spark documentation/UAT WIP. * UAT-based updates to parameterized expectations Signed-off-by: James Campbell <james.p.campbell@gmail.com> * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Self-Managed Spark Datasource HOWTO guide. * Docs/draft docs for data connectors (great-expectations#2086) Changes proposed in this pull request: This PR adds how-to guides for new-style DataConnectors and Datasources for 0.13. Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: William Shin <will@superconductive.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Update docs/guides/how_to_guides/configuring_datasources/how_to_configure_a_self_managed_spark_datasource.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * Enhance self_check of Datasource class. * Linting * ExecutionEngine Config [WIP] * ExecutionEngine Config [WIP] * Update links * Add docs tab for experimental API * Add admonition pointing user to docs for experimental API * Update docs/guides/how_to_guides/creating_and_editing_expectations/how_to_create_parameterized_expectations_super_fast.rst Co-authored-by: Abe Gong <abegong@users.noreply.github.com> * ExecutionEngine Config * Tests for more expressive Datasource.self_check() diagnostic. * Linting. * TypeError fix for numeric utility. * Additional test. Co-authored-by: James Campbell <james.p.campbell@gmail.com> Co-authored-by: Rob Lim <robert.m.lim@gmail.com> Co-authored-by: William Shin <will@superconductive.com> Co-authored-by: Abe Gong <abegong@users.noreply.github.com> Co-authored-by: Abe <abegong@gmail.com> Co-authored-by: gilpasternak35 <gilpasternak35@gmail.com> Co-authored-by: Eugene Mandel <eugene@superconductivehealth.com>
Describe the bug
If asking GE for a complete result when
expect_column_values_to_not_be_null
. GE will try to return a result which contains samples of the values which failed the expectation (if result is set toCOMPLETE
). It fails when one of the values is the new pandasNA
type (GE internals fails to convert it).To Reproduce
Error message:
Expected behavior
To return a result and not fail. Maybe with pd.NA or just replace it with a string interpretation of
pd.NA
or aNone
as the returned list of values (which failed).Environment (please complete the following information):
The text was updated successfully, but these errors were encountered: