Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RLlib] Allow for more than 2^31 policy timesteps. #11301

Merged

Conversation

sven1977
Copy link
Contributor

@sven1977 sven1977 commented Oct 9, 2020

RLlib currently crashes when [policy].global_timestep reaches 2^31 due to the respective tensor-types being of dtype=int32.
This PR fixes this issue.

Issue #10810

Closes #10810

Closes #10810

  • I've run scripts/format.sh to lint the changes in this PR.
  • I've included any doc changes needed for https://docs.ray.io/en/master/.
  • I've made sure the tests are passing. Note that there might be a few flaky tests, see the recent failures at https://flakey-tests.ray.io/
  • Testing Strategy
    • Unit tests
    • Release tests
    • This PR is not tested :(

@sven1977 sven1977 added the tests-ok The tagger certifies test failures are unrelated and assumes personal liability. label Oct 12, 2020
Copy link
Contributor

@ericl ericl left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice test!

@ericl ericl merged commit 8ea1bc5 into ray-project:master Oct 12, 2020
@sven1977 sven1977 deleted the issue_10810_go_beoynd_2pow31_timesteps branch January 18, 2021 13:04
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
tests-ok The tagger certifies test failures are unrelated and assumes personal liability.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[rllib] How to train beyond 2^31 timesteps?
2 participants