From 5171b9bd497cb07c3e7ab940230c97916ab29f6a Mon Sep 17 00:00:00 2001 From: Dannon Baker Date: Wed, 18 Jul 2018 09:40:46 -0400 Subject: [PATCH] A couple more typos. --- test/base/populators.py | 18 +++++++++--------- test/integration/test_scripts.py | 2 +- 2 files changed, 10 insertions(+), 10 deletions(-) diff --git a/test/base/populators.py b/test/base/populators.py index 2d06142f987d..5072d8093ae2 100644 --- a/test/base/populators.py +++ b/test/base/populators.py @@ -225,7 +225,7 @@ def _summarize_history(self, history_id): @contextlib.contextmanager def test_history(self, **kwds): - # TODO: In the future allow targetting a specfic history here + # TODO: In the future allow targeting a specific history here # and/or deleting everything in the resulting history when done. # These would be cool options for remote Galaxy test execution. try: @@ -765,7 +765,7 @@ def __create_payload_fetch(self, history_id, collection_type, **kwds): if isinstance(contents, list): for i, contents_level in enumerate(contents): - # If given a full colleciton definition pass as is. + # If given a full collection definition pass as is. if isinstance(contents_level, dict): elements.append(contents_level) continue @@ -825,9 +825,9 @@ def paste_content_to_simluated_files(contents): ) return payload - def wait_for_fetched_collection(self, fetch_repsonse): - self.dataset_populator.wait_for_job(fetch_repsonse["jobs"][0]["id"], assert_ok=True) - initial_dataset_collection = fetch_repsonse["outputs"][0] + def wait_for_fetched_collection(self, fetch_response): + self.dataset_populator.wait_for_job(fetch_response["jobs"][0]["id"], assert_ok=True) + initial_dataset_collection = fetch_response["outputs"][0] dataset_collection = self.dataset_populator.get_history_collection_details(initial_dataset_collection["history_id"], hid=initial_dataset_collection["hid"]) return dataset_collection @@ -947,12 +947,12 @@ def read_test_data(test_dict): if "name" in value: new_collection_kwds["name"] = value["name"] if collection_type == "list:paired": - fetch_repsonse = dataset_collection_populator.create_list_of_pairs_in_history(history_id, contents=elements, **new_collection_kwds).json() + fetch_response = dataset_collection_populator.create_list_of_pairs_in_history(history_id, contents=elements, **new_collection_kwds).json() elif collection_type == "list": - fetch_repsonse = dataset_collection_populator.create_list_in_history(history_id, contents=elements, direct_upload=True, **new_collection_kwds).json() + fetch_response = dataset_collection_populator.create_list_in_history(history_id, contents=elements, direct_upload=True, **new_collection_kwds).json() else: - fetch_repsonse = dataset_collection_populator.create_pair_in_history(history_id, contents=elements or None, direct_upload=True, **new_collection_kwds).json() - hdca = dataset_populator.ds_entry(fetch_repsonse["outputs"][0]) + fetch_response = dataset_collection_populator.create_pair_in_history(history_id, contents=elements or None, direct_upload=True, **new_collection_kwds).json() + hdca = dataset_populator.ds_entry(fetch_response["outputs"][0]) label_map[key] = hdca inputs[key] = hdca has_uploads = True diff --git a/test/integration/test_scripts.py b/test/integration/test_scripts.py index 6d8ee51fa5c6..abc7a1fd4365 100644 --- a/test/integration/test_scripts.py +++ b/test/integration/test_scripts.py @@ -166,7 +166,7 @@ def _skip_if_not_postgres(self): raise unittest.SkipTest("Test only valid for postgres") def _scripts_check_argparse_help(self, script): - # Test imports and argparse repsonse to --help with 0 exit code. + # Test imports and argparse response to --help with 0 exit code. output = self._scripts_check_output(script, ["--help"]) # Test -h, --help in printed output message. assert "-h, --help" in output