Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DISABLED test_ddp_checkpointing_weight_sharing_use_reentrant_False (__main__.TestDDPWithWorkWrapper) #103876

Closed
rohan-varma opened this issue Jun 20, 2023 · 2 comments
Labels
module: ci Related to continuous integration module: ddp Issues/PRs related distributed data parallel training skipped Denotes a (flaky) test currently skipped in CI. triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module

Comments

@rohan-varma
Copy link
Member

rohan-varma commented Jun 20, 2023

Reason

Broken by #103487. The pending fix is in #103629

cc @seemethere @malfet @pytorch/pytorch-dev-infra

@rohan-varma rohan-varma added the module: ci Related to continuous integration label Jun 20, 2023
@pytorch-bot pytorch-bot bot added the skipped Denotes a (flaky) test currently skipped in CI. label Jun 20, 2023
@pytorch-bot
Copy link

pytorch-bot bot commented Jun 20, 2023

Hello there! From the DISABLED prefix in this issue title, it looks like you are attempting to disable a test in PyTorch CI. The information I have parsed is below:
  • Test name: test_ddp_checkpointing_weight_sharing_use_reentrant_False (__main__.TestDDPWithWorkWrapper)
  • Platforms for which to skip the test: none parsed, defaulting to ALL platforms
  • Disabled by rohan-varma

Within ~15 minutes, test_ddp_checkpointing_weight_sharing_use_reentrant_False (__main__.TestDDPWithWorkWrapper) will be disabled in PyTorch CI for all platforms. Please verify that your test name looks correct, e.g., test_cuda_assert_async (__main__.TestCuda).

To modify the platforms list, please include a line in the issue body, like below. The default action will disable the test for all platforms if no platforms list is specified.

Platforms: case-insensitive, list, of, platforms

We currently support the following platforms: asan, dynamo, inductor, linux, mac, macos, rocm, slow, win, windows.

@janeyx99 janeyx99 added triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module module: ddp Issues/PRs related distributed data parallel training labels Jun 21, 2023
@pytorch-bot
Copy link

pytorch-bot bot commented Jul 17, 2023

Resolving the issue because the test is not flaky anymore after 150 reruns without any failures and the issue hasn't been updated in 14 days. Please reopen the issue to re-disable the test if you think this is a false positive

@pytorch-bot pytorch-bot bot closed this as completed Jul 17, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
module: ci Related to continuous integration module: ddp Issues/PRs related distributed data parallel training skipped Denotes a (flaky) test currently skipped in CI. triaged This issue has been looked at a team member, and triaged and prioritized into an appropriate module
Projects
None yet
Development

No branches or pull requests

2 participants