Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CI test linux://python/ray/dag:tests/experimental/test_torch_tensor_dag is consistently_failing #45923

Closed
can-anyscale opened this issue Jun 13, 2024 · 12 comments
Labels
bug Something that is supposed to be working; but isn't ci-test core Issues that should be addressed in Ray Core flaky-tracker Issue created via Flaky Test Tracker https://flaky-tests.ray.io/ ray-test-bot Issues managed by OSS test policy stability triage Needs triage (eg: priority, bug/not-bug, and owning component) weekly-release-blocker Issues that will be blocking Ray weekly releases

Comments

@can-anyscale
Copy link
Collaborator

CI test linux://python/ray/dag:tests/experimental/test_torch_tensor_dag is consistently_failing. Recent failures:
- https://buildkite.com/ray-project/postmerge/builds/4886#01901003-09a2-4538-a518-b77ce646c989
- https://buildkite.com/ray-project/postmerge/builds/4886#01901003-099e-41a2-869b-9a2e92c9c809

DataCaseName-linux://python/ray/dag:tests/experimental/test_torch_tensor_dag-END
Managed by OSS Test Policy

@can-anyscale can-anyscale added bug Something that is supposed to be working; but isn't ci-test core Issues that should be addressed in Ray Core flaky-tracker Issue created via Flaky Test Tracker https://flaky-tests.ray.io/ ray-test-bot Issues managed by OSS test policy stability triage Needs triage (eg: priority, bug/not-bug, and owning component) weekly-release-blocker Issues that will be blocking Ray weekly releases labels Jun 13, 2024
@can-anyscale
Copy link
Collaborator Author

Blamed commit: d577652 found by bisect job https://buildkite.com/ray-project/release-tests-bisect/builds/1231

@can-anyscale
Copy link
Collaborator Author

This test is now considered as flaky because it has been failing on postmerge for too long. Flaky tests do not run on premerge.

@can-anyscale
Copy link
Collaborator Author

blame reverted

@can-anyscale
Copy link
Collaborator Author

This test is now considered as flaky because it has been failing on postmerge for too long. Flaky tests do not run on premerge.

1 similar comment
@can-anyscale
Copy link
Collaborator Author

This test is now considered as flaky because it has been failing on postmerge for too long. Flaky tests do not run on premerge.

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

@can-anyscale
Copy link
Collaborator Author

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something that is supposed to be working; but isn't ci-test core Issues that should be addressed in Ray Core flaky-tracker Issue created via Flaky Test Tracker https://flaky-tests.ray.io/ ray-test-bot Issues managed by OSS test policy stability triage Needs triage (eg: priority, bug/not-bug, and owning component) weekly-release-blocker Issues that will be blocking Ray weekly releases
Projects
None yet
Development

No branches or pull requests

1 participant