Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Disable flaky twitter bots dataset loading test. #3439

Merged
merged 1 commit into from
Aug 11, 2023
Merged

Conversation

justinxzhao
Copy link
Collaborator

No description provided.

Comment on lines +18 to +27
# DISABLED: Flaky for tests, probably due to the dataset size.
# # Test loading dataset without 'split' and 'Unnamed: 0' columns in config.
# twitter_bots_config = ludwig.datasets._get_dataset_config("twitter_bots")
# assert isinstance(twitter_bots_config, DatasetConfig)

# twitter_bots_dataset = ludwig.datasets.get_dataset("twitter_bots", cache_dir=tmpdir)
# assert isinstance(twitter_bots_dataset, DatasetLoader)
# df = twitter_bots_dataset.load()
# assert df is not None
# assert len(df.columns) == 22 # Expected number of columns in Twitter bots dataset including split column.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should we skip the test? Or is this fine?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There's an upper half of the test that tests the basic mechanics of loading a dataset that hasn't been flake, so still want to keep the test overall.

The twitter bots block additionally checks that loading works when there's a pre-set split column, which seems incremental. My feeling is that it's not worth keeping it enabled since we already have the first test.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do you happen to remember what the purpose/intention of this test was? I remember we added this twitter bots component in the last year but I can't remember the reason for it. Just want to make sure it's ok to disable!

Going to temporarily approve for now

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks like this test was added in https://github.com/ludwig-ai/ludwig/pull/3285/files by @connor-mccorm to verify loading when there's a pre-set split column. Removing this test seems low risk to me, but it's certainly checking behavior that isn't covered by the first test.

Since the test hasn't failed for a few CI runs now, my guess is that the original test failed because of a momentary Kaggle outage, similar to how our HF tests used to fail whenever there was an HF outage.

I'd be more inclined to merge this PR if we've seen the test fail multiple times, but it's just been once, this PR could be an overreaction.

@github-actions
Copy link

Unit Test Results

  6 files  ±0    6 suites  ±0   1h 25m 3s ⏱️ + 3m 49s
33 tests ±0  29 ✔️ ±0    4 💤 ±0  0 ±0 
99 runs  ±0  87 ✔️ ±0  12 💤 ±0  0 ±0 

Results for commit 51d6891. ± Comparison against base commit 91c28f8.

@justinxzhao justinxzhao reopened this Aug 11, 2023
@justinxzhao
Copy link
Collaborator Author

Seems like this test has failed again in https://github.com/ludwig-ai/ludwig/actions/runs/5836017733/job/15828730546. Re-opening.

@justinxzhao justinxzhao merged commit 7ecdfa5 into master Aug 11, 2023
30 checks passed
@justinxzhao justinxzhao deleted the flaky_test branch August 11, 2023 20:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

2 participants