-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[rllib] move evaluation to trainer.step() such that the result is properly logged #12708
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks fine for me.
@Maltimore Could you fix the broken eval test case and run the ci/travis/format.sh (LINT) before pushing?
Thanks!
I ran
|
…gged even when training with trainer.train() instead of tune.run() lint bugfix: need to increment self._iteration in if condition [rllib] evaluation: simplify logic in if-condition
@sven1977 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for this fix @Maltimore !
Looks good to me.
…t is properly logged (ray-project#12708)" This reverts commit 59010e2.
…t is properly logged (ray-project#12708)" This reverts commit 59010e2.
Move evaluation to
trainer.step()
such that the result is properly logged even when training withtrainer.train()
instead oftune.run()
Why are these changes needed?
When training with
trainer.train()
instead oftune.run()
, the results of the evaluation (evaluation metrics) are not written to disk (e. g. to the progress.csv).This moves the evaluation code to
trainer.step()
, which ensures that the evaluation metrics are included in progress.csv and other output files.Related issue number
Checks
scripts/format.sh
to lint the changes in this PR. --> yes but this errors with a flake8 error unrelated to the codeI only tested this PR in my own experiments.