Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[DM-33438] Clean up run_demo.sh #31

Merged
merged 6 commits into from Jan 25, 2023
Merged

[DM-33438] Clean up run_demo.sh #31

merged 6 commits into from Jan 25, 2023

Conversation

cbanek
Copy link
Contributor

@cbanek cbanek commented Jan 24, 2023

No description provided.

@cbanek
Copy link
Contributor Author

cbanek commented Jan 24, 2023

This script works if I take out the multiprocessing stuff, but with the multiprocessing, I get the following error from the python tests:

___________________________________________________________________ PiplinesCheckTestCase.testExecutionButler (chain='demo_collection_exe') ___________________________________________________________________

self = <test_butler.PiplinesCheckTestCase testMethod=testExecutionButler>

def testExecutionButler(self):
    """Check outputs match in both runs."""

    for chain in (EXE_CHAIN, QBB_CHAIN):
        with self.subTest(chain=chain):
            # Check that we have identical datasets in both collections
            # except for the dataset.id
            main_datasets = self._get_datasets_from_chain(MAIN_CHAIN)
            datasets = self._get_datasets_from_chain(chain)
            self.assertGreater(len(datasets), 0)
          self.assertEqual(len(main_datasets), len(datasets))

E AssertionError: 21 != 13

tests/test_butler.py:101: AssertionError
___________________________________________________________________ PiplinesCheckTestCase.testExecutionButler (chain='demo_collection_qbb') ___________________________________________________________________

self = <test_butler.PiplinesCheckTestCase testMethod=testExecutionButler>

def testExecutionButler(self):
    """Check outputs match in both runs."""

    for chain in (EXE_CHAIN, QBB_CHAIN):
        with self.subTest(chain=chain):
            # Check that we have identical datasets in both collections
            # except for the dataset.id
            main_datasets = self._get_datasets_from_chain(MAIN_CHAIN)
            datasets = self._get_datasets_from_chain(chain)
            self.assertGreater(len(datasets), 0)
          self.assertEqual(len(main_datasets), len(datasets))

E AssertionError: 21 != 13

tests/test_butler.py:101: AssertionError

bin/run_demo.sh Outdated Show resolved Hide resolved
@timj
Copy link
Member

timj commented Jan 24, 2023

That test is saying that the number of datasets created in the EXECUTION BUTLER bit of code (the first 3 pipetask calls using $NODE) is equal to the number of datasets created by the QUNTUM BACKED BUTLER bit of code (the second set of 3 pipetask calls using $NODE).

Copy link
Member

@timj timj left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This looks good. There is a small possibility that the two parallel processes hitting sqlite would fall over each other but given that the second run is using the graph and not the sqlite at all and we only have two processes, I think this will almost certainly be fine. cc/ @andy-slac

bin/run_demo.sh Outdated Show resolved Hide resolved
bin/run_demo.sh Outdated Show resolved Hide resolved
bin/run_demo.sh Outdated Show resolved Hide resolved
bin/run_demo.sh Outdated Show resolved Hide resolved
@cbanek cbanek merged commit 6f36787 into main Jan 25, 2023
@cbanek cbanek deleted the tickets/DM-33438 branch January 25, 2023 19:31
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants