diff --git a/docs/changelog/changelog.rst b/docs/changelog/changelog.rst index 2de1045c80bc..7271af85b0ad 100644 --- a/docs/changelog/changelog.rst +++ b/docs/changelog/changelog.rst @@ -19,10 +19,8 @@ develop * Remove the "project new" option from the command line (since it is not implemented; users can only run "init" to create a new project). * Update type detection for bigquery based on driver changes in pybigquery driver 0.4.14. Added a warning for users who are running an older pybigquery driver * added execution tests to the NotebookRenderer to mitigate codegen risks -* Add option "persist", true by default, for SparkDFDataset to persist the DataFrame it is passed. This addresses #1133 -in a deeper way (thanks @tejsvirai for the robust debugging support and reproduction on spark). - - Disabling this option should *only* be done if the user has *already* externally persisted the DataFrame, or if the -dataset is too large to persist but *computations are guaranteed to be stable across jobs*. +* Add option "persist", true by default, for SparkDFDataset to persist the DataFrame it is passed. This addresses #1133 in a deeper way (thanks @tejsvirai for the robust debugging support and reproduction on spark). + * Disabling this option should *only* be done if the user has *already* externally persisted the DataFrame, or if the dataset is too large to persist but *computations are guaranteed to be stable across jobs*. * Enable passing dataset kwargs through datasource via dataset_options batch_kwarg. * Fix AttributeError when validating expectations from a JSON file * Data Docs: fix bug that was causing erratic scrolling behavior when table of contents contains many columns @@ -31,10 +29,8 @@ dataset is too large to persist but *computations are guaranteed to be stable ac 0.9.7 ----------------- * Update marshmallow dependency to >3. NOTE: as of this release, you MUST use marshamllow >3.0, which REQUIRES python 3. (`#1187 `_) @jcampbell - - Schema checking is now stricter for expectation suites, and data_asset_name must not be present as a top-level - key in expectation suite json. It is safe to remove. - - Similarly, datasource configuration must now adhere strictly to the required schema, including having any - required credentials stored in the "credentials" dictionary. + * Schema checking is now stricter for expectation suites, and data_asset_name must not be present as a top-level key in expectation suite json. It is safe to remove. + * Similarly, datasource configuration must now adhere strictly to the required schema, including having any required credentials stored in the "credentials" dictionary. * New beta CLI command: `tap new` that generates an executable python file to expedite deployments. (`#1193 `_) @Aylr * bugfix in TableBatchKwargsGenerator docs * Added feature maturity in README (`#1203 `_) @kyleaton @@ -78,8 +74,8 @@ dataset is too large to persist but *computations are guaranteed to be stable ac * Add support for transient table creation in snowflake (#1012) * Improve path support in TupleStoreBackend for better cross-platform compatibility * New features on `ExpecatationSuite` - - `.add_citation()` - - `get_citations()` + - `.add_citation()` + - `get_citations()` * `SampleExpectationsDatasetProfiler` now leaves a citation containing the original batch kwargs * `great_expectations suite edit` now uses batch_kwargs from citations if they exist * Bugfix :: suite edit notebooks no longer blow away the existing suite while loading a batch of data diff --git a/great_expectations/cli/datasource.py b/great_expectations/cli/datasource.py index 86ac572283b5..baf59539cb6c 100644 --- a/great_expectations/cli/datasource.py +++ b/great_expectations/cli/datasource.py @@ -155,14 +155,6 @@ def datasource_profile(datasource, generator_name, data_assets, profile_all_data if the number of data assets in the datasource exceeds the internally defined limit. If it does, it will prompt the user to either specify the list of data assets to profile or to profile all. If the limit is not exceeded, the profiler will profile all data assets in the datasource. - - :param datasource: name of the datasource to profile - :param data_assets: if this comma-separated list of data asset names is provided, only the specified data assets will be profiled - :param profile_all_data_assets: if provided, all data assets will be profiled - :param directory: - :param view: Open the docs in a browser - :param additional_batch_kwargs: Additional keyword arguments to be provided to get_batch when loading the data asset. - :return: """ cli_message("Warning - this is a BETA feature.") try: